<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>3DG: A Framework for Using Generative AI for Handling Sparse Learner Performance Data From Intelligent Tutoring Systems⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Liang Zhang</string-name>
          <email>lzhang13@memphis.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jionghao Lin</string-name>
          <email>jionghao@cmu.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Conrad Borchers</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Meng Cao</string-name>
          <email>mcao@memphis.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiangen Hu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Electrical and Computer Engineering, University of Memphis</institution>
          ,
          <addr-line>Memphis, TN 38152</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Psychology, University of Memphis</institution>
          ,
          <addr-line>Memphis, TN, 38152</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Human-Computer Interaction Institute, Carnegie Mellon University</institution>
          ,
          <addr-line>Pittsburgh, PA, 15213</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Institute for Intelligent Systems, University of Memphis</institution>
          ,
          <addr-line>Memphis, TN 38152</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Learning performance data (e.g., quiz scores and attempts) is significant for understanding learner engagement and knowledge mastery level. However, the learning performance data collected from Intelligent Tutoring Systems (ITSs) often sufers from sparsity, impacting the accuracy of learner modeling and knowledge assessments. To address this, we introduce the 3DG framework (3-Dimensional tensor for Densification and Generation), a novel approach combining tensor factorization with advanced generative models, including Generative Adversarial Network (GAN) and Generative Pre-trained Transformer (GPT), for enhanced data imputation and augmentation. The framework operates by first representing the data as a three-dimensional tensor, capturing dimensions of learners, questions, and attempts. It then densifies the data through tensor factorization and augments it using Generative AI models, tailored to individual learning patterns identified via clustering. Applied to data from an AutoTutor lesson by the Center for the Study of Adult Literacy (CSAL), the 3DG framework efectively generated scalable, personalized simulations of learning performance. Comparative analysis revealed GAN's superior reliability over GPT-4 in this context, underscoring its potential in addressing data sparsity challenges in ITSs and contributing to the advancement of personalized educational technology.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Learning Performance Data</kwd>
        <kwd>Data Sparsity</kwd>
        <kwd>Intelligent Tutoring System</kwd>
        <kwd>Generative Model</kwd>
        <kwd>Generative Adversarial Network</kwd>
        <kwd>Generative Pre-trained Transformer</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Intelligent Tutoring System (ITS) is a prototype of computer system designed to ofer
personalized and adaptive instructions through tracing and analyzing learning performance data such as
quiz scores and question attempts [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. However, during the interaction between learners
and ITS, learning performance data often exhibits data sparsity due to unexplored questions,
insuficient attempts to master knowledge, and lacking variability in learning patterns [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8 ref9">5, 6, 7, 8, 9</xref>
        ].
Data sparsity can lead to biased analysis and modeling of learning data. This is particularly
evident in the “Learner Model” component of ITS, which is crucial for tracking learning and
predicting performance of individual learners [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ]. Specifically, sparse performance data
can lead to skewed or overfitted Knowledge Tracing models in “Learner Model”, which impedes
accurately capturing learner knowledge states and may result in misleading predictions of
learning performance [
        <xref ref-type="bibr" rid="ref13 ref14 ref6 ref9">6, 9, 13, 14</xref>
        ]. The scarcity of learning performance data significantly
hampers the development of ITSs, particularly in cases where learners have not suficiently
engaged with certain instructional scenarios [
        <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
        ].
      </p>
      <p>
        Tackling data sparsity for ITSs presents a practical yet challenging research area. Informed
by the machine learning literature [
        <xref ref-type="bibr" rid="ref17 ref18 ref19 ref20">17, 18, 19, 20</xref>
        ], the issue of data sparsity can be addressed by
two principal ways: data imputation and data augmentation. Firstly, data imputation focuses
on filling the gaps in missing data to ensure a comprehensive dataset [
        <xref ref-type="bibr" rid="ref21 ref6 ref8">6, 8, 21</xref>
        ]. Secondly, data
augmentation aims to enrich and expand datasets where there are insuficient learning patterns,
thus ensuring robustness in analysis, modeling, and even potential testing tasks for ITSs [
        <xref ref-type="bibr" rid="ref22 ref9">9, 22</xref>
        ].
Currently, limited eforts have been made in the field of ITSs to systematically address these
data sparsity issues in learning performance data [
        <xref ref-type="bibr" rid="ref21 ref5 ref6">5, 6, 21</xref>
        ]. Driven by this, we propose the 3DG
(3-dimensions, Densification, and Generation) simulation framework, a systematic approach
leveraging generative models to handle sparse learning performance from ITS.
      </p>
      <p>
        The 3DG framework derived from its three core phases. In the first phase , a 3-dimensional
tensor is constructed to represent learning performance data, with dimensions corresponding
to learners, questions, and attempts. The second phase focuses on densifying the sparse tensor
by tensor factorization. The third phase entails the generation of learning performance data
based on generative models, tailored to the individual learning patterns of learners. The 3DG
framework integrates the multidimensional learner model with generative models to facilitate
scalable simulation sampling for individual learning patterns. The multidimensional learner
model in our framework is derived from the Tensor Factorization method, a widely-used
approach in predicting learner performance in many studies [
        <xref ref-type="bibr" rid="ref21 ref23 ref24 ref25">21, 23, 24, 25</xref>
        ]. Initially, learning
performance values are represented in a three-dimensional tensor encompassing dimensions
of learners, questions, and attempts. Specifically, learning performance indicators, such as
binary responses from learners at problem-solving step attempts (with correct answers denoted
as 1 and incorrect as 0), form the tensor entries, and they are arranged sequentially along
the question queue in the learning process and sorted by attempts in ascending order. This
constructed tensor exhibits data sparsity. Our study aims to perform data imputation and
augmentation on the sparse tensor. Mathematically, the tensor factorization method addresses
incomplete and missing performance values in factorization computations, serving as a form
of tensor completion typically used in data imputation [
        <xref ref-type="bibr" rid="ref21 ref26 ref27">21, 26, 27</xref>
        ]). Inspired by the recent
advancements of generative models [
        <xref ref-type="bibr" rid="ref28 ref29">28, 29</xref>
        ], which are capable of generating data based on
patterns learned during training and have revolutionized simulation methodologies to be more
lfexible and cost-efective, our study delves into exploring their potential of addressing the
data sparsity issue. We operate under the foundational assumption that, if learning patterns
can be identified within the multidimensional learner model, they can be efectively simulated
and generated using generative models, facilitating scalable data augmentation. Consequently,
current research was guided by following two Research Questions:
• RQ 1: What is the most efective method for integrating tensor factorization and
generative models to develop a systematic framework that proficiently imputes and augments
sparse learning performance data?
• RQ 2: In the context of simulating learning performance data, how do Generative
Adversarial Network (GAN) and Generative Pre-trained Transformer (GPT) models compare in
terms of efectiveness and accuracy?
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methods</title>
      <sec id="sec-2-1">
        <title>2.1. Dataset</title>
        <p>
          Our study investigated a dataset derived from the AutoTutor ITS, focusing on learning
performance in reading comprehension. This dataset originates from lessons developed for the
Center for the Study of Adult Literacy (CSAL) [
          <xref ref-type="bibr" rid="ref30 ref31">30, 31</xref>
          ], specifically the ’Cause and Efect’ lesson,
involving 118 participants. The lesson design incorporates three levels of question dificulty:
medium (M), easy (E), and hard (H). There are 9 medium-dificulty questions, 10 easy questions,
and 10 hard questions. Notably, the distribution of learners across these dificulty levels varies
within the lesson. Upon completing the medium dificulty level, learners are either advanced to
the hard level or redirected to the easy level, depending on their performance, thus providing a
tailored learning pathway.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. The Systematic Simulation Framework</title>
        <p>We propose a systematic simulation framework, 3DG, illustrated in Figure 1. This framework
begins by structuring the initial learning performance data, sourcing from real-world
learnerITS interactions, into a three-dimensional tensor by dimensions of learners, questions, and
attempts. As depicted in the sparse cube space in Figure 1, filled cubes represent recorded values
of learning performance, while transparent cubes indicate missing values. Tensor completion
(based on tensor factorization) is then utilized, converting the sparse tensor to a densified one.
The densified tensor provides invaluable information in identifying various learning patterns,
which aids in dividing the tensor into sub-tensors by categorizing distinct learning patterns.
Subsequently, generative models are harnessed to simulate additional data samples for enriching
the original dataset based on each specific learning pattern. The entire operation is encapsulated
for scalable simulation sampling and ultimately ofers a comprehensive dataset incorporating
both imputed and augmented data. This framework was developed to address RQ 1. More
detailed methods within this framework are described in the following subsections.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Tensor Completion for Data Imputation</title>
        <p>The three-dimensional tensor  , representing the learning process, is defined as  ∈
× ×  , where the  = (1, 2, 3, · · · , ) is the maximum number of learners,  =
(1, 2, 3, · · · , ) the maximum number of questions, and  = (1, 2, 3, · · · , ) the
maximum number of attempts. Each element   of  indicates the performance variable
of learner  on question  at the attempt  . For instance, in the CSAL AutoTutor context,
a binary variable   = {0, 1} is used, where 1 signifies a correct answer and 0 denotes an
incorrect one. We model the tenor  as a factorization of two lower dimensional components:
1) a learner latent matrix  of size  ×  ( represents the set of latent features in tensor
factorization), which captures learner-related latent features matrix/space (such as abilities and
learning-related features); and 2) a latent tensor  of size  ×  ×  , representing the learner
knowledge in terms of latent features during question attempts. The approximated tensor ˆ is
obtained by the following formula:
ˆ
 ≈  × 
(1)
where  can be interpreted as the latent feature space encapsulating learner-related efects,
reflecting characteristics such as individual abilities/features and learning preferences. On the
other hand, the tensor  represents the interaction between attempts and question-related
(knowledge acquisition) efects, adapting to various learner features.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Scalable Simulation based on Generative Models for Data Augmentation</title>
        <p>To answer RQ 2, we used two generative models, GAN (Generative Adversarial Network) and
GPT (Generative Pre-trained Transformer), to facilitate scalable simulations that are tailored to
individual learning patterns. According to [32, 33], GAN model is uniquely structured with a
dual-network architecture comprising a generator and a discriminator. This architecture enables
GAN model to excel in generating high-quality synthetic data. In comparison, GPT model is
distinguished by its use of transformer architecture, which empowers it to generate data that is
not only contextually relevant but also maintains a high degree of coherence [34, 35].</p>
        <p>Before initiating the simulations, we employ a clustering algorithm (i.e., K-means++) to
categorize individual learning patterns based on similarities in learners’ performance. The
learners-attempts matrix slice extracted from the ˆ , encapsulates the probability-based
knowledge states associated with the performance on the th question , for all  learners over 
attempts. In our analysis, we employ the “power law learning curve", a model widely recognized
in educational and training research [36, 37, 38], to fit the learning performance with increasing
attempts. In the power-law formula  = , the  represents the learning performance,
quantified as the probability of producing correct answers, and  is the number of opportunities
to practice a skill or attempt. The parameter  indicates the measurement of the learner’s initial
ability or prior knowledge, and  represents the learning rate at which the learner acquires
knowledge through practice. We employ K-means++ [39, 40] to cluster the distribution of two
model parameters ( and ), which assists in identifying distinct individual learning patterns.</p>
        <p>As illustrated in Figure 2, the architecture of Generative Adversarial Network (GAN) consists
of two distinct neural networks: the Generator and the Discriminator. The Generator, often a
type of neural network like a convolutional neural network (CNN), is designed to create synthetic
data samples. It is denoted as (· ). The Discriminator, typically another neural network which
can also be a CNN (though its structure may vary based on the specific application), is tasked with
evaluating whether the data samples are real (authentic data) or fabricated by the Generator.
It is denoted as (· ). In the process (Figure 2), the Generator starts with a noise sample,
usually drawn from a Gaussian distribution, which has dimensions compatible with the original
data distribution. This noise sample serves as the initial input for the Generator, resulting
in  . Both this   and the   are then fed into the
Discriminator (· ). The Discriminator’s role is to discern whether each sample is real or
simulated. Concurrently, the Generator (· ) is trained to progressively reduce the diference
between the distributions of the real and simulated data through iterative tuning.</p>
        <p>Considering the limitations of using purely numerical values for interoperability and the
enhanced semantic understanding that detailed descriptions provide, we have developed a
mixedbased prompt approach for GPT-4 (illustrated in Figure 3). The prompt strategy integrates
original matrix data with interpretive text, thereby enriching the context and interpretability
of the data. Additionally, it incorporates the Chain-of-Thought (CoT) prompting technique
[41], which involves appending guiding phrases such as ’Let’s think step by step’ at the end
of the prompt to facilitate a more structured analytical process. Specifically, the constructed
prompt includes comprehensive elements such as the reading material being analyzed, detailed
information about the questions (including their answers), and the learners-attempts matrix
data, complete with descriptive information about both its format and entries. Subsequently, a
simulation request prompts GPT-4 to integrate the numerical and textual data in a coherent and
insightful manner, ultimately driving the execution of a simulation. During the optimization
process, these prompts are iteratively refined and adjusted to eficiently yield results that align
with our specified objectives.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <p>
        As illustrated in Table 1, the original dataset exhibits sparsity levels ranging from 80% to 85%
(as determined by calculating the proportion of missing values to the total number of entries).
By iteratively tuning the latent feature range [
        <xref ref-type="bibr" rid="ref1 ref20">1, 20</xref>
        ] in tensor factorization algorithms, we
identified the optimal number of latent features ( ) as 6 for both Lesson (M) and Lesson (H),
and 4 for Lesson (E). The optimal  value was derived by averaging results from multiple
trainings with optimized  values in tensor factorization.
      </p>
      <p>These findings suggest that tensor completion (based on tensor factorization) can eficiently
impute missing values in the original sparse performance data, notably for unexplored questions
and attempts. This enhancement is crucial for facilitating more comprehensive analysis and
modeling in Intelligent Tutoring Systems (ITSs). The latent features, closely associated with
learner-specific characteristics during the learning process, are captured with nuanced detail,
particularly in the context of reading comprehension. Further research is imperative to fully
understand the underlying physical essence of these latent features.</p>
      <p>The distributions of parameters  and  are illustrated in Figure 4 and Figure 5, respectively.
These figures visualize the parameter distributions from a example cluster data set with an
original size of 20, and exhibit simulations in increments of 1000, with total sizes ranging from
1000 to 20000.</p>
      <p>Figure 4 demonstrates the distributions of parameters , which is used to represent the
learner’s initial ability or prior knowledge. Figure 4a shows the distribution of the parameter 
obtained by GAN simulation. As the sample size increases, the range of parameter  from the
simulation sample mostly falls within the original range of parameter , although it exhibits
a longer tail distribution extending beyond the original maximum value of parameter . The
distribution of the parameter  obtained by GPT-4 simulation is illustrated in Figure 4b. Unlike
those obtained from GAN simulation, the range of parameter  values here extends beyond
the original range, which is particularly evident as the simulated sample size increases. This
(a) Distribution of parameter  by GAN simulation.</p>
      <p>(b) Distribution of parameter  by GPT-4 simulation.
suggests that the initial learning ability in GPT-4 simulated samples exhibits more variability
and divergence from the original data compared to those from GAN simulation.</p>
      <p>Then, Figure 5 demonstrates the distributions of parameters  as derived from both GAN and
GPT-4 simulations. The parameter  represents the learning rate, which reflects how quickly
a learner acquires knowledge through practice. The GAN simulation produces a narrower
range of parameter  values, especially in terms of maximum and minimum values, when
compared to the original dataset, as depicted in Figure 5a. With increasing sample size, this
range generally maintains a consistent pattern. On the other hand, the GPT-4 simulation,
as shown in Figure 5b, demonstrates a broader range for parameter , extending beyond the
original scope. This contrast suggests that GPT-4 simulation may capture a wider variability in
learning rates compared to GAN simulation.</p>
      <p>(a) Distribution of parameter  by GAN simulation.</p>
      <p>(b) Distribution of parameter  by GPT-4 simulation.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion and Conclusion</title>
      <p>This paper proposed the 3DG systematic simulation framework based on generative models
(particularly GAN and GPT) to address data sparsity challenges in learning performance data
within intelligent tutoring systems (ITS). The framework involves representing learner data from
problem-solving step attempts as a three-dimensional tensor with the axes of learners, questions,
and attempts. Tensor completion, based on tensor factorization, is then utilized to impute missing
performance data entries, generating a dense tensor. Such imputation computation leverages
the similarities in learner performance across various questions and attempts, capturing the
sequential and temporal dynamics of learning [42, 43]. We have demonstrated the integration
of generative models, including GAN and GPT-4 for creating scalable, individualized learning
simulations aimed at enhancing learner models for personalized instruction. Our comparative
analysis reveals that GAN surpasses GPT-4 in terms of reliability for scalable simulations.</p>
      <p>Overall, the GAN simulations demonstrate a narrower and more consistent range of values
for parameters  and , indicating higher reliability for scalable simulations compared to
the broader value range exhibited by the GPT-4 simulations. The mechanism for the
GPT4 simulations, refined through iterative optimization of GPT-4 prompts, involves selecting
random values from a flat array of the original data. These values are then adjusted to match
base probabilities, preserving the overall data distribution while facilitating the creation of
an expanded dataset. Although valuable in computational simulations, this method generally
underperforms in numerical computing compared to deep learning models, as demonstrated by
the GAN’s performance in this study.</p>
      <p>Our findings shed light on the potential use of GPT-4 in simulating learner performance
represented through numerical values in future research. Firstly, employing mixed-based prompts
improves interoperability with numerical data, thus enhancing the eficiency of subsequent
modeling and simulation computations. Secondly, the Chain-of-Thought (CoT) prompting
technique delineates the steps for the simulation task, efectively directing GPT-4 in its reasoning
process. This includes a structured approach comprising: Understanding the Existing Matrix,
Distribution Analysis, Clustering Information, and Simulation Process. Thirdly, the
computational power of GPT-4 in modeling and simulation is attributed to its capabilities in
self-search, self-programming, and self-computing, all of which are facilitated by prompt
engineering. This significantly enhances its utility in data analysis and modeling for future research
endeavors. However, integrating GPT-4 with numerical computation presents fundamental
challenges, as we discuss in the following section.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Limitations and Future Works</title>
      <p>The capability of GPT-4 in performing deep learning tasks involving numerical computations
remains insuficient, primarily due to the intrinsic limitations of large language models and
platform contraints. Future research could productively explore the integration of GAN with GPT
models, aiming to improve their interoperability and computational capabilities. Furthermore,
the degree of sparsity in the original performance data, particularly when formatted as a
tensor, significantly impacts the performance of generative models. Therefore, investigating the
sensitivity and robustness of tensor completion methods in response to diferent levels of data
sparsity presents an important avenue for future studies. Such investigations are crucial for
better integrating large language models within Intelligent Tutoring Systems (ITSs), potentially
leading to more refined and efective educational tools.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgements</title>
      <p>We extend our sincere gratitude to Prof. Philip I. Pavlik Jr. from the University of Memphis and
Prof. Shaghayegh Sahebi from the University at Albany - SUNY for their expert guidance on
tensor factorization method.
P. Pavlik, et al., Clustering the learning patterns of adults with low literacy skills interacting
with an intelligent tutoring system., Grantee Submission (2018).
[32] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville,
Y. Bengio, Generative adversarial nets, Advances in neural information processing systems
27 (2014).
[33] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville,
Y. Bengio, Generative adversarial networks, Communications of the ACM 63 (2020)
139–144.
[34] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I.
Polosukhin, Attention is all you need, Advances in neural information processing systems 30
(2017).
[35] A. Radford, K. Narasimhan, T. Salimans, I. Sutskever, et al., Improving language
understanding by generative pre-training (2018).
[36] A. Newell, P. S. Rosenbloom, Mechanisms of skill acquisition and the law of practice.,
Technical Report, CARNEGIE-MELLON UNIV PITTSBURGH PA DEPT OF COMPUTER
SCIENCE, 1980.
[37] H. Cen, K. Koedinger, B. Junker, Learning factors analysis–a general method for cognitive
model evaluation and improvement, in: International conference on intelligent tutoring
systems, Springer, 2006, pp. 164–175.
[38] R. DeKeyser, Skill acquisition theory, in: Theories in second language acquisition,</p>
      <p>Routledge, 2020, pp. 83–104.
[39] D. Arthur, S. Vassilvitskii, K-means++ the advantages of careful seeding, in: Proceedings of
the eighteenth annual ACM-SIAM symposium on Discrete algorithms, 2007, pp. 1027–1035.
[40] B. Bahmani, B. Moseley, A. Vattani, R. Kumar, S. Vassilvitskii, Scalable k-means++, arXiv
preprint arXiv:1203.6402 (2012).
[41] J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou, et al.,
Chainof-thought prompting elicits reasoning in large language models, Advances in Neural
Information Processing Systems 35 (2022) 24824–24837.
[42] C. M. Conway, M. H. Christiansen, Sequential learning in non-human primates, Trends in
cognitive sciences 5 (2001) 539–546.
[43] C. M. Conway, Sequential Learning, Springer US, Boston, MA, 2012, pp. 3047–3050.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. F.</given-names>
            <surname>Boyle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Reiser</surname>
          </string-name>
          ,
          <source>Intelligent tutoring systems, Science</source>
          <volume>228</volume>
          (
          <year>1985</year>
          )
          <fpage>456</fpage>
          -
          <lpage>462</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Corbett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Koedinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <article-title>Intelligent tutoring systems, in: Handbook of human-computer interaction</article-title>
          , Elsevier,
          <year>1997</year>
          , pp.
          <fpage>849</fpage>
          -
          <lpage>874</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>K.</given-names>
            <surname>VanLehn,</surname>
          </string-name>
          <article-title>The relative efectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems</article-title>
          ,
          <source>Educational psychologist 46</source>
          (
          <year>2011</year>
          )
          <fpage>197</fpage>
          -
          <lpage>221</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Graesser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sottilare</surname>
          </string-name>
          ,
          <article-title>Intelligent tutoring systems, in: International handbook of the learning sciences</article-title>
          ,
          <source>Routledge</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>246</fpage>
          -
          <lpage>255</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.-N.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <article-title>Predicting student performance in an intelligent tutoring system</article-title>
          ,
          <source>Ph.D. thesis, Stiftung Universität Hildesheim</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Karypis</surname>
          </string-name>
          ,
          <article-title>A self-attentive model for knowledge tracing</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>06837</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Sun, Attention-based knowledge tracing with heterogeneous information network embedding, in: Knowledge Science, Engineering</article-title>
          and Management: 13th International Conference, KSEM 2020, Hangzhou, China,
          <source>August 28-30</source>
          ,
          <year>2020</year>
          , Proceedings,
          <source>Part I 13</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>95</fpage>
          -
          <lpage>103</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sahebi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Brusilovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. O.</given-names>
            <surname>Moraes</surname>
          </string-name>
          ,
          <article-title>Knowledge tracing for complex problem solving: Granular rank-based tensor factorization</article-title>
          ,
          <source>in: Proceedings of the 29th ACM Conference on User Modeling, Adaptation and Personalization</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>179</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cui</surname>
          </string-name>
          , L. Xu,
          <article-title>Graphca: Learning from graph counterfactual augmentation for knowledge tracing</article-title>
          ,
          <source>IEEE/CAA Journal of Automatica Sinica</source>
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <fpage>2108</fpage>
          -
          <lpage>2123</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P. I. Pavlik</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. G.</given-names>
            <surname>Eglington</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Zhang,</surname>
          </string-name>
          <article-title>Automatic domain model creation and improvement</article-title>
          .,
          <string-name>
            <surname>Grantee Submission</surname>
          </string-name>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L. G.</given-names>
            <surname>Eglington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. I. Pavlik</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <article-title>How to optimize student learning using student models that adapt rapidly to individual diferences</article-title>
          ,
          <source>International Journal of Artificial Intelligence in Education</source>
          <volume>33</volume>
          (
          <year>2023</year>
          )
          <fpage>497</fpage>
          -
          <lpage>518</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. I. Pavlik</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Cockroft</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Shi, Exploring the individual diferences in multidimensional evolution of knowledge states of learners</article-title>
          , in: International Conference on Human-Computer Interaction, Springer,
          <year>2023</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>284</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <article-title>Deep hierarchical knowledge tracing</article-title>
          ,
          <source>in: Proceedings of the 12th international conference on educational data mining</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>W.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Park</surname>
          </string-name>
          , S. Park,
          <article-title>Contrastive learning for knowledge tracing</article-title>
          ,
          <source>in: Proceedings of the ACM Web Conference</source>
          <year>2022</year>
          ,
          <year>2022</year>
          , pp.
          <fpage>2330</fpage>
          -
          <lpage>2338</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kossiakof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. N.</given-names>
            <surname>Sweet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Seymour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Biemer</surname>
          </string-name>
          ,
          <article-title>Systems engineering principles and practice</article-title>
          , volume
          <volume>83</volume>
          , John Wiley &amp; Sons,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Baudin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dutfoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Iooss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-L.</given-names>
            <surname>Popelin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Openturns:</surname>
          </string-name>
          <article-title>An industrial software for uncertainty quantification in simulation</article-title>
          ,
          <source>in: Handbook of uncertainty quantification</source>
          , Springer,
          <year>2017</year>
          , pp.
          <fpage>2001</fpage>
          -
          <lpage>2038</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>E.</given-names>
            <surname>Acar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Dunlavy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Kolda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mørup</surname>
          </string-name>
          ,
          <article-title>Scalable tensor factorizations for incomplete data</article-title>
          ,
          <source>Chemometrics and Intelligent Laboratory Systems</source>
          <volume>106</volume>
          (
          <year>2011</year>
          )
          <fpage>41</fpage>
          -
          <lpage>56</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C.</given-names>
            <surname>Shorten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Khoshgoftaar</surname>
          </string-name>
          ,
          <article-title>A survey on image data augmentation for deep learning</article-title>
          ,
          <source>Journal of big data 6</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>48</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>T.</given-names>
            <surname>Emmanuel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Maupong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mpoeleng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Semong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mphago</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Tabona</surname>
          </string-name>
          ,
          <article-title>A survey on missing data in machine learning</article-title>
          ,
          <source>Journal of Big Data</source>
          <volume>8</volume>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>T.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <article-title>Adaptive data augmentation for supervised learning over missing data</article-title>
          ,
          <source>Proceedings of the VLDB Endowment</source>
          <volume>14</volume>
          (
          <year>2021</year>
          )
          <fpage>1202</fpage>
          -
          <lpage>1214</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>N.</given-names>
            <surname>Thai-Nghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Drumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Horváth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Krohn-Grimberghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nanopoulos</surname>
          </string-name>
          , L. SchmidtThieme,
          <article-title>Factorization techniques for predicting student performance, in: Educational recommender systems and technologies: Practices and challenges</article-title>
          ,
          <source>IGI Global</source>
          ,
          <year>2012</year>
          , pp.
          <fpage>129</fpage>
          -
          <lpage>153</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          , I. Sutskever,
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <article-title>Imagenet classification with deep convolutional neural networks</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>25</volume>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sahebi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-R.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Brusilovsky</surname>
          </string-name>
          ,
          <article-title>Tensor factorization for student modeling and performance prediction in unstructured domain</article-title>
          .,
          <source>International Educational Data Mining Society</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>T.-N.</given-names>
            <surname>Doan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sahebi</surname>
          </string-name>
          ,
          <article-title>Rank-based tensor factorization for student performance prediction</article-title>
          ,
          <source>in: 12th International Conference on Educational Data Mining (EDM)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sahebi</surname>
          </string-name>
          ,
          <article-title>Modeling knowledge acquisition from multiple learning resource types</article-title>
          , arXiv preprint arXiv:
          <year>2006</year>
          .
          <volume>13390</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>N.</given-names>
            <surname>Thai-Nghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Horváth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          , et al.,
          <article-title>Factorization models for forecasting student performance</article-title>
          ., in: EDM, Eindhoven,
          <year>2011</year>
          , pp.
          <fpage>11</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Missing trafic data imputation and pattern discovery with a bayesian augmented tensor factorization model</article-title>
          ,
          <source>Transportation Research Part C: Emerging Technologies</source>
          <volume>104</volume>
          (
          <year>2019</year>
          )
          <fpage>66</fpage>
          -
          <lpage>77</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>M.</given-names>
            <surname>Jovanovic</surname>
          </string-name>
          , M. Campbell,
          <article-title>Generative artificial intelligence: Trends and prospects</article-title>
          ,
          <source>Computer</source>
          <volume>55</volume>
          (
          <year>2022</year>
          )
          <fpage>107</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>D.</given-names>
            <surname>Baidoo-Anu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Owusu</given-names>
            <surname>Ansah</surname>
          </string-name>
          ,
          <article-title>Education in the era of generative artificial intelligence (ai): Understanding the potential benefits of chatgpt in promoting teaching and learning</article-title>
          ,
          <source>Available at SSRN</source>
          <volume>4337484</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Graesser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. O.</given-names>
            <surname>Baer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Olney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Greenberg</surname>
          </string-name>
          ,
          <article-title>Reading comprehension lessons in autotutor for the center for the study of adult literacy, in: Adaptive educational technologies for literacy instruction</article-title>
          ,
          <source>Routledge</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>288</fpage>
          -
          <lpage>293</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shubeck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lippert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gatewood</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>