<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Deep learning Grade Prediction Model of Online Learning Performance Based on knowledge learning representation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Shuaileng</forename><surname>Yuan</surname></persName>
							<affiliation key="aff0">
								<orgName type="laboratory">Laboratory for Image and Media Understanding</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sukrit</forename><surname>Leelaluk</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Graduate School of Information Science and Electrical Engineering</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cheng</forename><surname>Tang</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Graduate School of Information Science and Electrical Engineering</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Li</forename><surname>Chen</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Graduate School of Information Science and Electrical Engineering</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Fumiya</forename><surname>Okubo</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Graduate School of Information Science and Electrical Engineering</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Atsushi</forename><surname>Shimada</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Graduate School of Information Science and Electrical Engineering</orgName>
								<orgName type="institution">Kyushu University</orgName>
								<address>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Deep learning Grade Prediction Model of Online Learning Performance Based on knowledge learning representation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6D665CCB471B8649896364F37E49C97F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:53+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Performance prediction</term>
					<term>Knowledge representation</term>
					<term>TabNet</term>
					<term>Behavior representation １</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In recent years, due to the impact of Coronavirus disease (COVID-19), digital platforms have developed rapidly and accumulated a large amount of data. To better utilize the comprehensive and diverse data stored in online platforms for data mining, such as learning behavior analysis or performance prediction, and to provide guidance and valuable feedback for educator became more important. For the current analysis of learning behaviors by time series data with DNN method, the interpretability is not enough. This paper proposes a method based on the simultaneous use of learning behaviors and learning materials to obtain the representation of learned knowledge, and through multiple cross-validations, the effect of this knowledge representation has a certain improvement on the original data, and the interpretability can promote the feedback function.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>With the continuous development of school online digital platform construction, students have generated a lot of data of learning life, but the data accumulated in the system as an asset has not yet fully released its value. Making good use of educational data resources with the help of data mining technology could provide scientific guidance and valuable decision-making support for school teaching, student learning and other tasks <ref type="bibr" target="#b0">[1]</ref>. It is unable to intervene in learners' bad learning behaviors in time and cannot provide help when some students have poor learning results. If we can predict that students are at risk of failing a class, and analyze learning method through data mining, better support can be used to guarantee learning result <ref type="bibr" target="#b1">[2]</ref>.</p><p>At present, many research methods focus on handling relational model of learning behavior by machine learning <ref type="bibr" target="#b2">[3]</ref>[3] <ref type="bibr" target="#b3">[4]</ref>, finding key influencing factors through correlation analysis of behavioral statistical data, or using NLP model with the time series data of behavior type or text <ref type="bibr" target="#b5">[6]</ref>. A major challenge is that correlation analysis is difficult to find complex relationships, and the DNN model that processes time series data has weak interpretability. Based on a relatively intuitive understanding, we can see that the process of students interacting with teaching materials during the learning process records a process of students acquiring knowledge. Here we proposed a method to predict the grade by learning behavior and corresponding knowledge, which can obtain richer information and interpretability at the same time.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Study performance individual difference</head><p>The modern humanistic point of view pays more attention to the development of learners' learning attitude and learning motivation in the learning process. understanding learners' characteristics and learning willingness to meet learners' learning interests and learning resources <ref type="bibr" target="#b6">[7]</ref>. According to humanistic behavioral theory, through the analysis of student behavior to judge the result, one of the key points is to obtain the development process of the student's learning motivation and learning attitude from the student behavior data <ref type="bibr" target="#b8">[8]</ref>. The method we propose is to better reflect these differences in feature representation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">TabNet</head><p>Students' performance is recorded as tabular data, and there were some widely used DNN model for this kind of data. TabNet <ref type="bibr" target="#b9">[9]</ref> is designed, which not only absorbs the advantages of the tree model, but also inherits the advantages of DNN. TabNet has been designed based on retaining the end-to-end and representation learning characteristics of DNN. It also has the advantages of tree model and sparse feature selection, which makes it suitable for interpretability. We choose TabNet DNN structure to process tabular data of learning behaviors, which can automatically mine learning method pattern information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.1.">Feature selecting process</head><p>TabNet has a decision-tree function as the Figure <ref type="figure" target="#fig_0">1</ref> : TabNet also uses the idea of sequential attention to imitate the behavior of decision trees. Each step of TabNet's decision-making module will allocate attention through mask function and obtain feature information by selecting features that have a greater impact on the results. Its interpretability can be defined by the formula below. 𝑀 !,# of each decision can be analyzed as pattern information, and the final 𝑀 $%%&amp;!,# can be used as the overall attention attributes for correlation analysis and other functions.</p><formula xml:id="formula_0">𝑀 $%%&amp;!,# = ∑ 𝜂 ! [𝑖]𝑀 !,# [𝑖]</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Study aim</head><p>The purpose of this study is based on the data of the existing digital learning platform, to establish an effective system that combines course materials and corresponding learning records to predict students' learning results. Providing well interpretability at the same time. In this proposed system, our goal is to explore the relationship between knowledge of different topics and learning performance in vectorized learning materials and establish a connected module to help obtain the feature representation of learning results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Proposed method</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Structure of proposed model</head><p>The focus of this study is to combine learning materials and learning performance to establish an effective system to predict student learning outcomes, providing a reference for the analysis of teaching and student learning situations.</p><p>From the system overview in Figure <ref type="figure" target="#fig_1">2</ref>, we can see that the system is divided into two parts: the knowledge representation part and the learning performance part. Each student's learning materials and online learning operation behavior records are used as input, and the final score is used as a label. The knowledge representation part is mainly responsible for converting learning materials into knowledge vectors. These knowledge vectors contain some connections and correlations between various knowledge points of learning. The learning performance part is mainly responsible for converting the operation records of the student's learning process into feature vectors of the learning performance of learning this part of knowledge. Then, by exploring the effect of the learning results on the knowledge vector, a knowledge vector representing the learning results is obtained, and performance predictions are made based on this. The output is the probability of each student achieving different grades. Because there are few direct connections between material from different courses, experiments can only be conducted for the same course. During training, the model records the data of some students in the same course as a training set and evaluates the effectiveness of the model by predicting the performance of the remaining students. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Preprocessing slide material</head><p>A typical teaching slide, which briefly expresses the main knowledge points of the course content, such as the meaning of some concepts, described formulas and relationships, etc. Some parts will also describe knowledge points through some pictures. Here we use pure text parts to summarize the main knowledge points representing the teaching materials. Mainly from the perspective of specific nouns and the relationships between them, we explore the internal relationships and interrelationships between knowledge points on different topics. When making textbooks, the main knowledge points in different chapters are somewhat different, but at the same time they are also connected according to certain correlations. Therefore, under normal circumstances, the internal relationship and correlation of knowledge points in different chapters are reflected to a certain extent through the distribution of nouns related to the knowledge points. So here we use all the text data in the course slides, first perform certain basic processing, and using Doc2Vec to convert it into a vector expression of word embedding. These word embeddings express information about the structure and content of a certain degree of course knowledge. Therefore, it can reflect the differences in the distribution ratio and structure of knowledge points in different chapters, which could also be regarded as sets of knowledge vector.</p><p>The original data like Figure <ref type="figure" target="#fig_2">3</ref> mainly includes the specific course chapter information of each course, the start and end time of each chapter, the student ID of each course, the event stream records of all students operating on the online platform, the operation part mainly including events such as turning pages forward and backward, clicking on links, annotations, notes, etc <ref type="bibr" target="#b10">[10]</ref> <ref type="bibr" target="#b11">[11]</ref>. Here we structured data through the API OpenLA <ref type="bibr" target="#b12">[12]</ref> which is developed by our research laboratory. The original data generated by the learning platform is the information generated by all operations saved based on time, including the username of the operation, time, operation type, etc., and using OpenLA, we can obtain the structured data we need according to specific needs. As you can see from the Figure <ref type="figure">4</ref>, what we use here is the statistical record of the operational behaviors of all students during weekly study. Including forward, backward, add bookmarks, jump and so on.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">The concept of knowledge vectors</head><p>As mentioned above that the meaningful words of each chapter of the course can represent information about the content and structure of a certain degree of knowledge. On the other hand, the learning process is the process of integrating new knowledge into one's own knowledge structure, which is the mastery of the content of knowledge itself and the correlation between knowledge. Therefore, based on these analyses, it is considered here that knowledge (descriptive words) embedding vector can be used as a standard to measure the effect of learning.</p><p>The materials studied by all students are consistent, that is, the knowledge content being learned is consistent. However, when different students study, their mastery of different knowledge points will be different, and the establishment of relationships between knowledge will also be different, so in the end there will be many differences in the learning results under the same learning materials. From this perspective, analyzing students' learning performance based on learning results based on knowledge content has a certain theoretical basis and is also consistent with some qualitative analysis results.</p><p>Figure <ref type="figure">4</ref>: The study process from knowledge vector perspective From the Figure <ref type="figure">4</ref>, we can see that if we look at the entire learning process from the perspective of knowledge vectors, the learning process is like a process of moving forward in the knowledge space which is containing all the knowledge vector of the course. In the beginning, it is at the origin of this knowledge space, which means not mastering any knowledge, after each period of study, students will reach a new position in the space. The boundary represents that certain dimensions reach an appropriate range, and students can get a good understanding of the course and pass the course.</p><p>When embedding the knowledge vector, the dimension should be considered with the amount of information of the text. More dimension means more complex relationship of knowledge. For not enough words in the slide, we set the dimension to 100.</p><p>The materials in each chapter are consistent, that is, the knowledge vectors being learned are consistent. However, if the influence of learning ability and learning habits is mapped on the knowledge vector, based on the standard vector that is being learned, different proportions will be obtained in different dimensions. Specifically from the right part of Figure <ref type="figure">4</ref>, for example, user 1 and user 2 here have very different learning modes, so the direction and distance of learning the same chapter in the knowledge space are quite different, but in the end, a similar total learning effect has been achieved; and user 1 and user 3 have similar learning patterns, so they have similar directions in each study, but due to different learning abilities, time spent and other reasons, the positions they end up are completely different, and the final results are naturally different.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Feature encoding for learning performance</head><p>In the past, data analysis of learning behavior was generally mainly combined with manual analysis. Selecting the part that we need from the various information existing in the learning operation data and combine it with other analysis models based on learning behavior for verification.</p><p>The main advantages of using TabNet to analyze learning behavior lie in two points. Firstly, no preprocessing is required, reducing manual operations in the process, and providing a more convenient way of use for the overall system design; Secondly, it can automatically display the data in tabular form the ability to automatically select features and combination relationships from the data. That means it could automatically search for student behavior patterns and behavioral rules in records, and automatically discover students' implicit learning motivations and learning habits and other information.</p><p>The specific process is to use students' operational data as input and student scores as labels for training. Finally, after training, the encoder structure of the trained TabNet is used to convert the student's operation data into a dense vector that meets the usage requirements. We can see by Figure <ref type="figure" target="#fig_3">5</ref>, TabNet model will take multiple steps. The process of each step is like the decisionmaking process of a decision tree. Different attentions are used to obtain learning performance information with different pattern information. The vectors obtained by the encoder in all steps are combined into a new vector. This new vector contains learning performance information based on different patterns, so it also contains richer pattern information of the learning process . </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Reflecting learning process on knowledge vector space</head><p>As mentioned earlier, the knowledge vector represented by word embedding contains the structural relationship of the learned knowledge on different topics, and the knowledge vectors of different topics have different emphases in different dimensions like Figure <ref type="figure">6</ref>. The knowledge vectors of some chapters will be larger in some dimensions. proportion. Therefore, we can speculate that if the knowledge of certain topics is better learned during the actual learning process, the results in these dimensions will also be better.</p><p>Figure <ref type="figure">6</ref>: the learning contents have different effects among dimensions Therefore, we can try to explore the impact of performance feature on specific parts of knowledge, and these different knowledges have different emphases in different dimensions. So here we can try to translate it into the impact of focusing on different dimensions of the knowledge space. The method we propose here is to convert the characteristics of learning performance into weights of different dimensions in the knowledge space, and then multiply the weights and the corresponding knowledge vectors in each dimension to obtain the final learned knowledge vector.  <ref type="figure" target="#fig_4">7</ref> Taking the learning performance feature 𝑃 ( as input, and finally the output of this MLP has the same dimension as the knowledge vector. This output is the effect of learning performance in different dimensions on the knowledge vector, that is the weight value 𝑊 ( (𝑤 * , 𝑤 , , … , 𝑤 -). The better the learning effect, the closer this effect weight value will be to 1, and the final knowledge vector will be closer to the standard value. By multiplying the output of MLP and the knowledge vector in one-to-one correspondence in each dimension, the vector expression 𝑆 ( of the learned knowledge is obtained. This output would be transferred through another fully connected layer, and the probability of the final score of this knowledge vector is obtained through SoftMax. The better the training effect, the better the effect of M1 on reflecting the influence of learning feature vectors to knowledge vectors.</p><formula xml:id="formula_1">𝑆 ( = 𝐷 ( U 𝑊 (<label>(2)</label></formula><formula xml:id="formula_2">𝑃𝑜𝑟𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 = 𝑠𝑜𝑓𝑡𝑚𝑎𝑥?𝐹𝐶(𝑆 ( )B<label>(3)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiment</head><p>To evaluate the effectiveness of the designed model, here we conducted a series of experiments to verify our method. The experiment used data of BookRoll reading behavior datasets and all slide text in a CS course, recording the log data of all student operation events. These mainly include operational events such as turning pages forward and backward, clicking links, marking, and taking notes. Through 5-fold cross-validation, training sets and validation sets are randomly selected from groups with different scores in different proportions. At the same time, in a more practical way, according to the standard of whether they are in the danger zone, students with different grades are finally classified into two categories to test the effect of the model. Because in the overall data, there is a gap of nearly 10 times between students in the danger zone and non-danger zone, so when setting the task indicators, students in the danger zone are set as label 1 to mainly examine the performance of predicting in risk crisis. It can more effectively identify students who need to provide early warning. The second experiment is using same prediction model, inputting knowledge representation of our method and behavior data separately, to verify whether our method extract more meaningful information for students with different behavior.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Result</head><p>To effectively evaluate the prediction quality of this knowledge vector model, here we selected traditional machine learning algorithms commonly used in other similar studies in the past as comparison baselines, including logistic regression (LR), extreme gradient boosting (XGBOOST), Lightgbm, random forest. The algorithm of knowledge vector proposed here have a certain improvement than the result predicted by behavior data alone. Performance improvements have been achieved in both F1 score and AUROC. Result shows our knowledge representation is more efficient than behavioral data only, which has ability to explore individual differences such as different grade of similar behaviors, or same grade of different behavior. It is worth mentioning that when the model we proposed predicts the learned knowledge vector, it only uses a simple fully connected layer to predict the results, which shows that the knowledge representation has a better classification boundary when predicting multiple grade categories. At risk students who were consistently mispredicted using behavioral data alone were more likely to be correctly identified by our method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Pattern information of TabNet</head><p>From Figure <ref type="figure" target="#fig_0">11</ref> we can see that for the most important three features in step1 and step2 of TabNet's decision steps, the coefficient of variation of students with good learning effects is larger, which means that the features discovered by TabNet are meaningful. If we conduct a qualitative analysis of "NOTGETIT, BOOKMARKJUMP, SEARCHJUMP", we can guess that the explanation of this pattern information is: For students with strong learning initiative, they will be more proactive in learning difficult-to-understand knowledge points after marking them. They may learn relevant content by searching. Therefore, this learning behavior pattern may correspond to the difficult knowledge in each chapter. So, we provide feedback on students' learning behavior and help improve learning results by counting the records of better performance in these patterns and higher performance records of their own. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper we hope to predict student learning result by combining the course material data of the digital education platform and the operation record data of student learning process. By introducing the idea of learning knowledge expression, we can dig out the result expression of learning knowledge from the materials to be learned and the learning process records. Combined with TabNet's advantages of not requiring preprocessing and interpretability on tabular data, it can provide better help for subsequent analysis and feedback work. It can be seen from the experimental results that the prediction model established here provides good prediction results.</p><p>As the current structure is not suitable for different courses' grade predicting. For more different courses, it is necessary to establish a set of methods to effectively handle the knowledge expression of all courses together. Based on the current excellent performance of the BERT model in large-scale text expression, in the future we hope to explore further using the BERT model to adapt to prediction methods for different courses in terms of knowledge expression.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Embodying the decision-making process</figDesc><graphic coords="2,119.35,425.95,370.80,175.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The structure of proposed method</figDesc><graphic coords="3,67.80,494.10,464.40,209.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: the original data of event stream and the structured data by OpenLA</figDesc><graphic coords="4,86.20,344.07,214.85,113.64" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: the multiple decision making step of TabNet</figDesc><graphic coords="6,128.48,161.38,352.55,208.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: the training process of the network</figDesc><graphic coords="7,86.20,162.27,388.18,260.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Comparison of at risk student prediction</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: The possible pattern information discovered by TabNet</figDesc><graphic coords="9,86.20,358.28,374.91,196.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="5,72.00,110.69,458.55,164.40" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The effect of online review exercises on student course engagement and learning performance: A case study of an introductory financial accounting course at an international joint venture university</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ding</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Accounting Education</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page">100699</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note>J</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Modeling and predicting students&apos; academic performance using data mining techniques</title>
		<author>
			<persName><forename type="first">A</forename><surname>Mueen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zafar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Manzoor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Modern Education and Computer Science</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page">36</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note>J</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The Learning Behavior Analysis of Online Vocational Education Students and Learning Resource Recommendation Based on Big Data</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">.</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Emerging Technologies in Learning (iJET)</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="issue">20</biblScope>
			<biblScope unit="page" from="261" to="273" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Reliable deep grade prediction with uncertainty estimation</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Rangwala</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th International Conference on Learning Analytics &amp; Knowledge</title>
				<meeting>the 9th International Conference on Learning Analytics &amp; Knowledge</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="76" to="85" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Prediction model on student performance based on internal assessment using deep learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hussain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Muhsin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Salal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Emerging Technologies in Learning</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">8</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note>J</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">LSTM: Efficient Student Modeling for knowledge tracing and student performance prediction</title>
		<author>
			<persName><forename type="first">S</forename><surname>Minn</surname></persName>
		</author>
		<author>
			<persName><surname>Bkt-</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2012.12218</idno>
	</analytic>
	<monogr>
		<title level="j">J</title>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Bouton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Context, attention, and the switch between habit and goal-direction in behavior</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title/>
	</analytic>
	<monogr>
		<title level="j">Learning &amp; behavior</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="349" to="362" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Analysis of Habit Patterns in Academic Behavior in Student Learning Discussions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Zulfa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Setiawan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Fardani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Elementary Education</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="392" to="399" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note>J</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">TabNet: Attentive interpretable tabular learning[C</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Ö</forename><surname>Arik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pfister</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI conference on artificial intelligence</title>
				<meeting>the AAAI conference on artificial intelligence</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="6679" to="6687" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Learning Analytics Platform in Higher Education in Japan</title>
		<author>
			<persName><forename type="first">B</forename><surname>Flanagan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ogata</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge Management &amp; E-Learning (KM&amp;EL)</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="469" to="484" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A Quality Data Set for Data Challenge: Featuring 160 Students&apos; Learning Behaviors and Learning Strategies in a Programming Course</title>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">H</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Flanagan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ogata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">30th International Conference on Computers in Education Conference Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="64" to="73" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">OpenLA: Library for Efficient E-book Log Analysis and Accelerating Learning Analytics</title>
		<author>
			<persName><forename type="first">R</forename><surname>Murata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Minematsu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shimada</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Computer in Education (ICCE 2020)</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="301" to="306" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
