<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">3DG: A Framework for Using Generative AI for Handling Sparse Learner Performance Data From Intelligent Tutoring Systems ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Liang</forename><surname>Zhang</surname></persName>
							<email>lzhang13@memphis.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Institute for Intelligent Systems</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Department of Electrical and Computer Engineering</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jionghao</forename><surname>Lin</surname></persName>
							<email>jionghao@cmu.edu</email>
							<affiliation key="aff2">
								<orgName type="department">Human-Computer Interaction Institute</orgName>
								<orgName type="institution">Carnegie Mellon University</orgName>
								<address>
									<postCode>15213</postCode>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Conrad</forename><surname>Borchers</surname></persName>
							<email>cborcher@cs.cmu.edu</email>
							<affiliation key="aff2">
								<orgName type="department">Human-Computer Interaction Institute</orgName>
								<orgName type="institution">Carnegie Mellon University</orgName>
								<address>
									<postCode>15213</postCode>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Meng</forename><surname>Cao</surname></persName>
							<email>mcao@memphis.edu</email>
							<affiliation key="aff3">
								<orgName type="department">Department of Psychology</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xiangen</forename><surname>Hu</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Institute for Intelligent Systems</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Department of Electrical and Computer Engineering</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="department">Department of Psychology</orgName>
								<orgName type="institution">University of Memphis</orgName>
								<address>
									<postCode>38152</postCode>
									<settlement>Memphis</settlement>
									<region>TN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">3DG: A Framework for Using Generative AI for Handling Sparse Learner Performance Data From Intelligent Tutoring Systems ⋆</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">5B24E7DF9D1D0D2949BB4F0341FE0A31</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:52+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Learning Performance Data, Data Sparsity, Intelligent Tutoring System, Generative Model, Generative Adversarial Network, Generative Pre-trained Transformer X. Hu) 0009-0002-0017-2569 (L. Zhang)</term>
					<term>0000-0003-3320-3907 (J. Lin)</term>
					<term>0000-0003-3437-8979 (C. Borchers)</term>
					<term>0000-0002-1286-2885 (M. Cao)</term>
					<term>0000-0001-9045-4070 (X. Hu)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Learning performance data (e.g., quiz scores and attempts) is significant for understanding learner engagement and knowledge mastery level. However, the learning performance data collected from Intelligent Tutoring Systems (ITSs) often suffers from sparsity, impacting the accuracy of learner modeling and knowledge assessments. To address this, we introduce the 3DG framework (3-Dimensional tensor for Densification and Generation), a novel approach combining tensor factorization with advanced generative models, including Generative Adversarial Network (GAN) and Generative Pre-trained Transformer (GPT), for enhanced data imputation and augmentation. The framework operates by first representing the data as a three-dimensional tensor, capturing dimensions of learners, questions, and attempts. It then densifies the data through tensor factorization and augments it using Generative AI models, tailored to individual learning patterns identified via clustering. Applied to data from an AutoTutor lesson by the Center for the Study of Adult Literacy (CSAL), the 3DG framework effectively generated scalable, personalized simulations of learning performance. Comparative analysis revealed GAN's superior reliability over GPT-4 in this context, underscoring its potential in addressing data sparsity challenges in ITSs and contributing to the advancement of personalized educational technology.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Intelligent Tutoring System (ITS) is a prototype of computer system designed to offer personalized and adaptive instructions through tracing and analyzing learning performance data such as quiz scores and question attempts <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref>. However, during the interaction between learners and ITS, learning performance data often exhibits data sparsity due to unexplored questions, insufficient attempts to master knowledge, and lacking variability in learning patterns <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref>. Data sparsity can lead to biased analysis and modeling of learning data. This is particularly evident in the "Learner Model" component of ITS, which is crucial for tracking learning and predicting performance of individual learners <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12]</ref>. Specifically, sparse performance data can lead to skewed or overfitted Knowledge Tracing models in "Learner Model", which impedes accurately capturing learner knowledge states and may result in misleading predictions of learning performance <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14]</ref>. The scarcity of learning performance data significantly hampers the development of ITSs, particularly in cases where learners have not sufficiently engaged with certain instructional scenarios <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref>.</p><p>Tackling data sparsity for ITSs presents a practical yet challenging research area. Informed by the machine learning literature <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b19">20]</ref>, the issue of data sparsity can be addressed by two principal ways: data imputation and data augmentation. Firstly, data imputation focuses on filling the gaps in missing data to ensure a comprehensive dataset <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b20">21]</ref>. Secondly, data augmentation aims to enrich and expand datasets where there are insufficient learning patterns, thus ensuring robustness in analysis, modeling, and even potential testing tasks for ITSs <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b21">22]</ref>. Currently, limited efforts have been made in the field of ITSs to systematically address these data sparsity issues in learning performance data <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b20">21]</ref>. Driven by this, we propose the 3DG (3-dimensions, Densification, and Generation) simulation framework, a systematic approach leveraging generative models to handle sparse learning performance from ITS.</p><p>The 3DG framework derived from its three core phases. In the first phase, a 3-dimensional tensor is constructed to represent learning performance data, with dimensions corresponding to learners, questions, and attempts. The second phase focuses on densifying the sparse tensor by tensor factorization. The third phase entails the generation of learning performance data based on generative models, tailored to the individual learning patterns of learners. The 3DG framework integrates the multidimensional learner model with generative models to facilitate scalable simulation sampling for individual learning patterns. The multidimensional learner model in our framework is derived from the Tensor Factorization method, a widely-used approach in predicting learner performance in many studies <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25]</ref>. Initially, learning performance values are represented in a three-dimensional tensor encompassing dimensions of learners, questions, and attempts. Specifically, learning performance indicators, such as binary responses from learners at problem-solving step attempts (with correct answers denoted as 1 and incorrect as 0), form the tensor entries, and they are arranged sequentially along the question queue in the learning process and sorted by attempts in ascending order. This constructed tensor exhibits data sparsity. Our study aims to perform data imputation and augmentation on the sparse tensor. Mathematically, the tensor factorization method addresses incomplete and missing performance values in factorization computations, serving as a form of tensor completion typically used in data imputation <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b25">26,</ref><ref type="bibr" target="#b26">27]</ref>). Inspired by the recent advancements of generative models <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29]</ref>, which are capable of generating data based on patterns learned during training and have revolutionized simulation methodologies to be more flexible and cost-effective, our study delves into exploring their potential of addressing the data sparsity issue. We operate under the foundational assumption that, if learning patterns can be identified within the multidimensional learner model, they can be effectively simulated and generated using generative models, facilitating scalable data augmentation. Consequently, current research was guided by following two Research Questions:</p><p>• RQ 1: What is the most effective method for integrating tensor factorization and generative models to develop a systematic framework that proficiently imputes and augments sparse learning performance data? • RQ 2: In the context of simulating learning performance data, how do Generative Adversarial Network (GAN) and Generative Pre-trained Transformer (GPT) models compare in terms of effectiveness and accuracy?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Dataset</head><p>Our study investigated a dataset derived from the AutoTutor ITS, focusing on learning performance in reading comprehension. This dataset originates from lessons developed for the Center for the Study of Adult Literacy (CSAL) <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b30">31]</ref>, specifically the 'Cause and Effect' lesson, involving 118 participants. The lesson design incorporates three levels of question difficulty: medium (M), easy (E), and hard (H). There are 9 medium-difficulty questions, 10 easy questions, and 10 hard questions. Notably, the distribution of learners across these difficulty levels varies within the lesson. Upon completing the medium difficulty level, learners are either advanced to the hard level or redirected to the easy level, depending on their performance, thus providing a tailored learning pathway.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">The Systematic Simulation Framework</head><p>We propose a systematic simulation framework, 3DG, illustrated in Figure <ref type="figure" target="#fig_0">1</ref>. This framework begins by structuring the initial learning performance data, sourcing from real-world learner-ITS interactions, into a three-dimensional tensor by dimensions of learners, questions, and attempts. As depicted in the sparse cube space in Figure <ref type="figure" target="#fig_0">1</ref>, filled cubes represent recorded values of learning performance, while transparent cubes indicate missing values. Tensor completion (based on tensor factorization) is then utilized, converting the sparse tensor to a densified one.</p><p>The densified tensor provides invaluable information in identifying various learning patterns, which aids in dividing the tensor into sub-tensors by categorizing distinct learning patterns. Subsequently, generative models are harnessed to simulate additional data samples for enriching the original dataset based on each specific learning pattern. The entire operation is encapsulated for scalable simulation sampling and ultimately offers a comprehensive dataset incorporating both imputed and augmented data. This framework was developed to address RQ 1. More detailed methods within this framework are described in the following subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Tensor Completion for Data Imputation</head><p>The three-dimensional tensor 𝒯 , representing the learning process, is defined as 𝒯 ∈ 𝑅 𝑈 ×𝑁 ×𝑀 , where the </p><formula xml:id="formula_0">𝑈 = 𝑚𝑎𝑥(1, 2, 3, • • • , 𝑢) is the maximum number of learners, 𝑁 = 𝑚𝑎𝑥(1, 2, 3, • • • , 𝑛) the</formula><formula xml:id="formula_1">𝒯 ˆ≈ 𝒰 × 𝒱<label>(1)</label></formula><p>where 𝒰 can be interpreted as the latent feature space encapsulating learner-related effects, reflecting characteristics such as individual abilities/features and learning preferences. On the other hand, the tensor 𝒱 represents the interaction between attempts and question-related (knowledge acquisition) effects, adapting to various learner features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Scalable Simulation based on Generative Models for Data Augmentation</head><p>To answer RQ 2, we used two generative models, GAN (Generative Adversarial Network) and GPT (Generative Pre-trained Transformer), to facilitate scalable simulations that are tailored to individual learning patterns. According to <ref type="bibr" target="#b31">[32,</ref><ref type="bibr" target="#b32">33]</ref>, GAN model is uniquely structured with a dual-network architecture comprising a generator and a discriminator. This architecture enables GAN model to excel in generating high-quality synthetic data. In comparison, GPT model is distinguished by its use of transformer architecture, which empowers it to generate data that is not only contextually relevant but also maintains a high degree of coherence <ref type="bibr" target="#b33">[34,</ref><ref type="bibr" target="#b34">35]</ref>. Before initiating the simulations, we employ a clustering algorithm (i.e., K-means++) to categorize individual learning patterns based on similarities in learners' performance. The learners-attempts matrix slice extracted from the 𝒯 ˆ, encapsulates the probability-based knowledge states associated with the performance on the 𝑛th question 𝑞 𝑛 , for all 𝑈 learners over 𝑀 attempts. In our analysis, we employ the "power law learning curve", a model widely recognized in educational and training research <ref type="bibr" target="#b35">[36,</ref><ref type="bibr" target="#b36">37,</ref><ref type="bibr" target="#b37">38]</ref>, to fit the learning performance with increasing attempts. In the power-law formula 𝑌 = 𝑎𝑋 𝑏 , the 𝑌 represents the learning performance, quantified as the probability of producing correct answers, and 𝑋 is the number of opportunities to practice a skill or attempt. The parameter 𝑎 indicates the measurement of the learner's initial ability or prior knowledge, and 𝑏 represents the learning rate at which the learner acquires knowledge through practice. We employ K-means++ <ref type="bibr" target="#b38">[39,</ref><ref type="bibr" target="#b39">40]</ref> to cluster the distribution of two model parameters (𝑎 and 𝑏), which assists in identifying distinct individual learning patterns.</p><p>As illustrated in Figure <ref type="figure" target="#fig_1">2</ref>, the architecture of Generative Adversarial Network (GAN) consists of two distinct neural networks: the Generator and the Discriminator. The Generator, often a type of neural network like a convolutional neural network (CNN), is designed to create synthetic data samples. It is denoted as 𝐺(•). The Discriminator, typically another neural network which can also be a CNN (though its structure may vary based on the specific application), is tasked with evaluating whether the data samples are real (authentic data) or fabricated by the Generator. It is denoted as 𝐷(•). In the process (Figure <ref type="figure" target="#fig_1">2</ref>), the Generator starts with a noise sample, usually drawn from a Gaussian distribution, which has dimensions compatible with the original data distribution. This noise sample serves as the initial input for the Generator, resulting in 𝑆𝑖𝑚𝑢𝑙𝑎𝑡𝑒𝑑 𝐷𝑎𝑡𝑎. Both this 𝑆𝑖𝑚𝑢𝑙𝑎𝑡𝑒𝑑 𝐷𝑎𝑡𝑎 and the 𝑅𝑒𝑎𝑙 𝐷𝑎𝑡𝑎 are then fed into the Discriminator 𝐷(•). The Discriminator's role is to discern whether each sample is real or simulated. Concurrently, the Generator 𝐺(•) is trained to progressively reduce the difference between the distributions of the real and simulated data through iterative tuning.</p><p>Considering the limitations of using purely numerical values for interoperability and the enhanced semantic understanding that detailed descriptions provide, we have developed a mixedbased prompt approach for GPT-4 (illustrated in Figure <ref type="figure" target="#fig_2">3</ref>). The prompt strategy integrates original matrix data with interpretive text, thereby enriching the context and interpretability of the data. Additionally, it incorporates the Chain-of-Thought (CoT) prompting technique <ref type="bibr" target="#b40">[41]</ref>, which involves appending guiding phrases such as 'Let's think step by step' at the end of the prompt to facilitate a more structured analytical process. Specifically, the constructed prompt includes comprehensive elements such as the reading material being analyzed, detailed information about the questions (including their answers), and the learners-attempts matrix data, complete with descriptive information about both its format and entries. Subsequently, a simulation request prompts GPT-4 to integrate the numerical and textual data in a coherent and  insightful manner, ultimately driving the execution of a simulation. During the optimization process, these prompts are iteratively refined and adjusted to efficiently yield results that align with our specified objectives.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>As illustrated in Table <ref type="table" target="#tab_1">1</ref>, the original dataset exhibits sparsity levels ranging from 80% to 85% (as determined by calculating the proportion of missing values to the total number of entries). By iteratively tuning the latent feature range <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b19">20]</ref> in tensor factorization algorithms, we identified the optimal number of latent features (𝐾) as 6 for both Lesson (M) and Lesson (H), and 4 for Lesson (E). The optimal 𝐾 value was derived by averaging results from multiple trainings with optimized 𝐾 values in tensor factorization.</p><p>These findings suggest that tensor completion (based on tensor factorization) can efficiently impute missing values in the original sparse performance data, notably for unexplored questions and attempts. This enhancement is crucial for facilitating more comprehensive analysis and modeling in Intelligent Tutoring Systems (ITSs). The latent features, closely associated with learner-specific characteristics during the learning process, are captured with nuanced detail, particularly in the context of reading comprehension. Further research is imperative to fully understand the underlying physical essence of these latent features. The distributions of parameters 𝑎 and 𝑏 are illustrated in Figure <ref type="figure" target="#fig_4">4</ref> and Figure <ref type="figure" target="#fig_6">5</ref>, respectively. These figures visualize the parameter distributions from a example cluster data set with an original size of 20, and exhibit simulations in increments of 1000, with total sizes ranging from 1000 to 20000.</p><p>Figure <ref type="figure" target="#fig_4">4</ref> demonstrates the distributions of parameters 𝑎, which is used to represent the learner's initial ability or prior knowledge. Figure <ref type="figure" target="#fig_4">4a</ref> shows the distribution of the parameter 𝑎 obtained by GAN simulation. As the sample size increases, the range of parameter 𝑎 from the simulation sample mostly falls within the original range of parameter 𝑎, although it exhibits a longer tail distribution extending beyond the original maximum value of parameter 𝑎. The distribution of the parameter 𝑎 obtained by GPT-4 simulation is illustrated in Figure <ref type="figure" target="#fig_4">4b</ref>. Unlike those obtained from GAN simulation, the range of parameter 𝑎 values here extends beyond the original range, which is particularly evident as the simulated sample size increases. This  suggests that the initial learning ability in GPT-4 simulated samples exhibits more variability and divergence from the original data compared to those from GAN simulation.</p><p>Then, Figure <ref type="figure" target="#fig_6">5</ref> demonstrates the distributions of parameters 𝑏 as derived from both GAN and GPT-4 simulations. The parameter 𝑏 represents the learning rate, which reflects how quickly a learner acquires knowledge through practice. The GAN simulation produces a narrower range of parameter 𝑏 values, especially in terms of maximum and minimum values, when compared to the original dataset, as depicted in Figure <ref type="figure" target="#fig_6">5a</ref>. With increasing sample size, this range generally maintains a consistent pattern. On the other hand, the GPT-4 simulation, as shown in Figure <ref type="figure" target="#fig_6">5b</ref>, demonstrates a broader range for parameter 𝑏, extending beyond the original scope. This contrast suggests that GPT-4 simulation may capture a wider variability in learning rates compared to GAN simulation.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion and Conclusion</head><p>This paper proposed the 3DG systematic simulation framework based on generative models (particularly GAN and GPT) to address data sparsity challenges in learning performance data within intelligent tutoring systems (ITS). The framework involves representing learner data from problem-solving step attempts as a three-dimensional tensor with the axes of learners, questions, and attempts. Tensor completion, based on tensor factorization, is then utilized to impute missing performance data entries, generating a dense tensor. Such imputation computation leverages the similarities in learner performance across various questions and attempts, capturing the sequential and temporal dynamics of learning <ref type="bibr" target="#b41">[42,</ref><ref type="bibr" target="#b42">43]</ref>. We have demonstrated the integration of generative models, including GAN and GPT-4 for creating scalable, individualized learning simulations aimed at enhancing learner models for personalized instruction. Our comparative analysis reveals that GAN surpasses GPT-4 in terms of reliability for scalable simulations.</p><p>Overall, the GAN simulations demonstrate a narrower and more consistent range of values for parameters 𝑎 and 𝑏, indicating higher reliability for scalable simulations compared to the broader value range exhibited by the GPT-4 simulations. The mechanism for the GPT-4 simulations, refined through iterative optimization of GPT-4 prompts, involves selecting random values from a flat array of the original data. These values are then adjusted to match base probabilities, preserving the overall data distribution while facilitating the creation of an expanded dataset. Although valuable in computational simulations, this method generally underperforms in numerical computing compared to deep learning models, as demonstrated by the GAN's performance in this study.</p><p>Our findings shed light on the potential use of GPT-4 in simulating learner performance represented through numerical values in future research. Firstly, employing mixed-based prompts improves interoperability with numerical data, thus enhancing the efficiency of subsequent modeling and simulation computations. Secondly, the Chain-of-Thought (CoT) prompting technique delineates the steps for the simulation task, effectively directing GPT-4 in its reasoning process. This includes a structured approach comprising: Understanding the Existing Matrix, Distribution Analysis, Clustering Information, and Simulation Process. Thirdly, the computational power of GPT-4 in modeling and simulation is attributed to its capabilities in self-search, self-programming, and self-computing, all of which are facilitated by prompt engineering. This significantly enhances its utility in data analysis and modeling for future research endeavors. However, integrating GPT-4 with numerical computation presents fundamental challenges, as we discuss in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Limitations and Future Works</head><p>The capability of GPT-4 in performing deep learning tasks involving numerical computations remains insufficient, primarily due to the intrinsic limitations of large language models and platform contraints. Future research could productively explore the integration of GAN with GPT models, aiming to improve their interoperability and computational capabilities. Furthermore, the degree of sparsity in the original performance data, particularly when formatted as a tensor, significantly impacts the performance of generative models. Therefore, investigating the sensitivity and robustness of tensor completion methods in response to different levels of data sparsity presents an important avenue for future studies. Such investigations are crucial for better integrating large language models within Intelligent Tutoring Systems (ITSs), potentially leading to more refined and effective educational tools.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The 3-dimensions, Densification, and Generation (3DG) systematic simulation framework.</figDesc><graphic coords="3,89.29,478.66,416.69,162.47" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Diagram of using generative adversarial network (GAN) model for data simulation.</figDesc><graphic coords="5,110.13,510.57,395.87,139.06" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Diagram of using generative pre-trained transformer (GPT) model for data simulation.</figDesc><graphic coords="6,99.71,84.19,395.86,174.12" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head></head><label></label><figDesc>(a) Distribution of parameter 𝑎 by GAN simulation. (b) Distribution of parameter 𝑎 by GPT-4 simulation.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Distribution of parameter 𝑎 by simulation.</figDesc><graphic coords="7,89.29,474.72,432.00,162.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>(a) Distribution of parameter 𝑏 by GAN simulation. (b) Distribution of parameter 𝑏 by GPT-4 simulation.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Distribution of parameter 𝑏 by simulation.</figDesc><graphic coords="8,89.29,426.10,432.00,162.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>maximum number of questions, and 𝑀 = 𝑚𝑎𝑥(1, 2, 3, • • • , 𝑚) the maximum number of attempts. Each element 𝜏 𝑢𝑖𝑗 of 𝒯 indicates the performance variable of learner 𝑙 𝑢 on question 𝑞 𝑖 at the attempt 𝑎 𝑗 . For instance, in the CSAL AutoTutor context, a binary variable 𝜏</figDesc><table /><note>𝑢𝑖𝑗 = {0, 1} is used, where 1 signifies a correct answer and 0 denotes an incorrect one. We model the tenor 𝒯 as a factorization of two lower dimensional components: 1) a learner latent matrix 𝒰 of size 𝑈 × 𝐾 (𝐾 represents the set of latent features in tensor factorization), which captures learner-related latent features matrix/space (such as abilities and learning-related features); and 2) a latent tensor 𝒱 of size 𝐾 × 𝑀 × 𝑁 , representing the learner knowledge in terms of latent features during question attempts. The approximated tensor 𝒯 ˆis obtained by the following formula:</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Results about the sparsity levels and latent features by tensor completion. M, H, and E denote Medium, Hard, and Easy lesson levels, respectively.</figDesc><table><row><cell>Dataset</cell><cell cols="2">Sparsity Level (Original) 𝐾 (Latent Features)</cell></row><row><cell>Lesson (M)</cell><cell>84.02%</cell><cell>6</cell></row><row><cell>Lesson (H)</cell><cell>85.45%</cell><cell>6</cell></row><row><cell>Lesson (E)</cell><cell>81.25%</cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Acknowledgements</head><p>We extend our sincere gratitude to Prof. Philip I. Pavlik Jr. from the University of Memphis and Prof. Shaghayegh Sahebi from the University at Albany -SUNY for their expert guidance on tensor factorization method.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Intelligent tutoring systems</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Anderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">F</forename><surname>Boyle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">J</forename><surname>Reiser</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Science</title>
		<imprint>
			<biblScope unit="volume">228</biblScope>
			<biblScope unit="page" from="456" to="462" />
			<date type="published" when="1985">1985</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Intelligent tutoring systems</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Corbett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Koedinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Anderson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Handbook of human-computer interaction</title>
				<imprint>
			<publisher>Elsevier</publisher>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="849" to="874" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The relative effectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems</title>
		<author>
			<persName><forename type="first">K</forename><surname>Vanlehn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Educational psychologist</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="197" to="221" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Intelligent tutoring systems</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Graesser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sottilare</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International handbook of the learning sciences</title>
				<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="246" to="255" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Predicting student performance in an intelligent tutoring system</title>
		<author>
			<persName><forename type="first">T.-N</forename><surname>Nguyen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
		<respStmt>
			<orgName>Stiftung Universität Hildesheim</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Pandey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Karypis</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.06837</idno>
		<title level="m">A self-attentive model for knowledge tracing</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Attention-based knowledge tracing with heterogeneous information network embedding</title>
		<author>
			<persName><forename type="first">N</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Science, Engineering and Management: 13th International Conference, KSEM 2020</title>
				<meeting><address><addrLine>Hangzhou, China</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">August 28-30, 2020. 2020</date>
			<biblScope unit="page" from="95" to="103" />
		</imprint>
	</monogr>
	<note>Proceedings, Part I 13</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Knowledge tracing for complex problem solving: Granular rank-based tensor factorization</title>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sahebi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Brusilovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">O</forename><surname>Moraes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th ACM Conference on User Modeling, Adaptation and Personalization</title>
				<meeting>the 29th ACM Conference on User Modeling, Adaptation and Personalization</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="179" to="188" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Graphca: Learning from graph counterfactual augmentation for knowledge tracing</title>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE/CAA Journal of Automatica Sinica</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="2108" to="2123" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Automatic domain model creation and improvement</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">I</forename><surname>Pavlik</surname><genName>Jr</genName></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">G</forename><surname>Eglington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">Grantee Submission</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">How to optimize student learning using student models that adapt rapidly to individual differences</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">G</forename><surname>Eglington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">I</forename><surname>Pavlik</surname><genName>Jr</genName></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Artificial Intelligence in Education</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="497" to="518" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Exploring the individual differences in multidimensional evolution of knowledge states of learners</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">I</forename><surname>Pavlik</surname><genName>Jr</genName></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Cockroft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Shi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Human-Computer Interaction</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="265" to="284" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Deep hierarchical knowledge tracing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th international conference on educational data mining</title>
				<meeting>the 12th international conference on educational data mining</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Contrastive learning for knowledge tracing</title>
		<author>
			<persName><forename type="first">W</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Park</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACM Web Conference 2022</title>
				<meeting>the ACM Web Conference 2022</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="2330" to="2338" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Kossiakoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">N</forename><surname>Sweet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Seymour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Biemer</surname></persName>
		</author>
		<title level="m">Systems engineering principles and practice</title>
				<imprint>
			<publisher>John Wiley &amp; Sons</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="volume">83</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Openturns: An industrial software for uncertainty quantification in simulation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Baudin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dutfoy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Iooss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-L</forename><surname>Popelin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Handbook of uncertainty quantification</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="2001" to="2038" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Scalable tensor factorizations for incomplete data</title>
		<author>
			<persName><forename type="first">E</forename><surname>Acar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Dunlavy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">G</forename><surname>Kolda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mørup</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Chemometrics and Intelligent Laboratory Systems</title>
		<imprint>
			<biblScope unit="volume">106</biblScope>
			<biblScope unit="page" from="41" to="56" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A survey on image data augmentation for deep learning</title>
		<author>
			<persName><forename type="first">C</forename><surname>Shorten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Khoshgoftaar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of big data</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="1" to="48" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A survey on missing data in machine learning</title>
		<author>
			<persName><forename type="first">T</forename><surname>Emmanuel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Maupong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mpoeleng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Semong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mphago</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Tabona</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Big Data</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="1" to="37" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Adaptive data augmentation for supervised learning over missing data</title>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Du</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the VLDB Endowment</title>
				<meeting>the VLDB Endowment</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="1202" to="1214" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Factorization techniques for predicting student performance</title>
		<author>
			<persName><forename type="first">N</forename><surname>Thai-Nghe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Drumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Horváth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Krohn-Grimberghe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nanopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Educational recommender systems and technologies: Practices and challenges</title>
				<imprint>
			<publisher>IGI Global</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="129" to="153" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Imagenet classification with deep convolutional neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Krizhevsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Tensor factorization for student modeling and performance prediction in unstructured domain</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sahebi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>-R. Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Brusilovsky</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>International Educational Data Mining Society</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Rank-based tensor factorization for student performance prediction</title>
		<author>
			<persName><forename type="first">T.-N</forename><surname>Doan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sahebi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">12th International Conference on Educational Data Mining (EDM)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sahebi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2006.13390</idno>
		<title level="m">Modeling knowledge acquisition from multiple learning resource types</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Factorization models for forecasting student performance</title>
		<author>
			<persName><forename type="first">N</forename><surname>Thai-Nghe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Horváth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">EDM</title>
				<meeting><address><addrLine>Eindhoven</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="11" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Missing traffic data imputation and pattern discovery with a bayesian augmented tensor factorization model</title>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transportation Research Part C: Emerging Technologies</title>
		<imprint>
			<biblScope unit="volume">104</biblScope>
			<biblScope unit="page" from="66" to="77" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Generative artificial intelligence: Trends and prospects</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jovanovic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Campbell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="107" to="112" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Education in the era of generative artificial intelligence (ai): Understanding the potential benefits of chatgpt in promoting teaching and learning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Baidoo-Anu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">Owusu</forename><surname>Ansah</surname></persName>
		</author>
		<idno>SSRN 4337484</idno>
	</analytic>
	<monogr>
		<title level="j">Available</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Reading comprehension lessons in autotutor for the center for the study of adult literacy</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Graesser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">O</forename><surname>Baer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Olney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Reed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Greenberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Adaptive educational technologies for literacy instruction</title>
				<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="288" to="293" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<title level="m" type="main">Clustering the learning patterns of adults with low literacy skills interacting with an intelligent tutoring system</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shubeck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lippert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gatewood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pavlik</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">Grantee Submission</note>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Generative adversarial nets</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pouget-Abadie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Warde-Farley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ozair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Generative adversarial networks</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pouget-Abadie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Warde-Farley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ozair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">63</biblScope>
			<biblScope unit="page" from="139" to="144" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ł</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<title level="m" type="main">Improving language understanding by generative pre-training</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Narasimhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Salimans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m" type="main">Mechanisms of skill acquisition and the law of practice</title>
		<author>
			<persName><forename type="first">A</forename><surname>Newell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Rosenbloom</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1980">1980</date>
			<publisher>CARNEGIE-MELLON UNIV PITTSBURGH PA DEPT OF COMPUTER SCIENCE</publisher>
		</imprint>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Learning factors analysis-a general method for cognitive model evaluation and improvement</title>
		<author>
			<persName><forename type="first">H</forename><surname>Cen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Koedinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Junker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on intelligent tutoring systems</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="164" to="175" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Skill acquisition theory</title>
		<author>
			<persName><forename type="first">R</forename><surname>Dekeyser</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Theories in second language acquisition</title>
				<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="83" to="104" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">K-means++ the advantages of careful seeding</title>
		<author>
			<persName><forename type="first">D</forename><surname>Arthur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vassilvitskii</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms</title>
				<meeting>the eighteenth annual ACM-SIAM symposium on Discrete algorithms</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="1027" to="1035" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Bahmani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Moseley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vattani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vassilvitskii</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1203.6402</idno>
		<title level="m">Scalable k-means++</title>
				<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">Chainof-thought prompting elicits reasoning in large language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="24824" to="24837" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">Sequential learning in non-human primates</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Conway</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">H</forename><surname>Christiansen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Trends in cognitive sciences</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="539" to="546" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<monogr>
		<title level="m" type="main">Sequential Learning</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Conway</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Springer US</publisher>
			<biblScope unit="page" from="3047" to="3050" />
			<pubPlace>Boston, MA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
