=Paper= {{Paper |id=Vol-2876/paper3 |storemode=property |title=Modelling Student Behavior in Synchronous Online Learning during the COVID-19 Pandemic |pdfUrl=https://ceur-ws.org/Vol-2876/paper3.pdf |volume=Vol-2876 |authors=Gianni Fenu,Roberta Galici |dblpUrl=https://dblp.org/rec/conf/wsdm/FenuG21 }} ==Modelling Student Behavior in Synchronous Online Learning during the COVID-19 Pandemic == https://ceur-ws.org/Vol-2876/paper3.pdf
Modelling Student Behavior in Synchronous Online
Learning during the COVID-19 Pandemic
Gianni Fenu, Roberta Galici
Department of Mathematics and Computer Science, University of Cagliari, V. Ospedale 72, 09124 Cagliari, Italy


                                      Abstract
                                      The outbreak of COVID-19 has significantly impacted on education, training, and mobility opportunities
                                      provided to learners, teachers, and educators. In response to this situation, synchronous online learn-
                                      ing has been massively adopted in universities and, therefore, the analysis of student behavior in this
                                      context is becoming essential. However, the literature has mainly tackled student behavior modelling in
                                      asynchronous online learning (e.g., interactions with pre-recorded videos), making how students learn
                                      in synchronous online learning so far under-explored. Grounding on the experience on online learning
                                      at the University of Cagliari, this paper proposes a preliminary study on student participation in syn-
                                      chronous lessons, covering more than 80 courses, and identifies patterns and strategies of engagement
                                      in classes over the semester. Then, by means of clustering techniques applied to data from more than
                                      25,000 students, we model fine-grained behavioral strategies and discuss how they are related with the
                                      student’s experience across courses and whether planning elements (e.g., the hour of the day a lesson
                                      is delivered) influence their level of participation. We expect that this study will support the educa-
                                      tional stakeholders with preliminary data-driven informed decisions and pose the basis for data-driven
                                      personalization for students and teachers involved in online synchronous learning.

                                      Keywords
                                      E-Learning, Engagement, Clustering, Learning Behavior, Student Modelling, Synchronous Learning.




1. Introduction
The Covid-19 pandemic has forced schools and universities to move face-to-face learning and
teaching in classrooms to online environments. The term online learning refers to educational
strategies that rely on the use of multimedia and Internet technologies for learning and teach-
ing. For instance, this strategy has been originally adopted to support students who cannot
physically attend lessons in classroom. Recent literature has focused on providing guidelines
on and discussing benefits and limitations of online learning, mainly identifying asynchronous
and synchronous strategies and addressing questions such as when, why and how to use these
two educational delivering modes [1]. Specifically, synchronous learning refers to learning ex-
periences where a group of participants is engaged at the same time, either in the same physical
location (e.g., a classroom) or the same online environment (e.g., a web conference room), and
can interact among each other in real time. Conversely, asynchronous learning refers to the


L2D’21: First International Workshop on Enabling Data-Driven Decisions from Learning on the Web, March 12, 2021,
Jerusalem, IL
" fenu@unica.it (G. Fenu); roberta.galici@unica.it (R. Galici)
 0000-0003-4668-2476 (G. Fenu); 0000-0002-1492-3514 (R. Galici)
                                    © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
 CEUR
 Workshop
 Proceedings
               http://ceur-ws.org
               ISSN 1613-0073       CEUR Workshop Proceedings (CEUR-WS.org)
strategies where instructors and students are not engaged in the learning process at the same
time (e.g., interactions with pre-recorded videos or on-demand online exams) [2].
   Given the increasing attention on synchronous and asynchronous online learning during
the last years [3, 4], analyzing how students behave in these contexts becomes of paramount
importance for several aspects, such as improving teaching strategies, sizing technological in-
frastructures, assessing and predicting student engagement, preventing students from losing
interest and possibly dropping out, and finally improving the student’s learning outcomes. It
has been proved that asynchronous learning benefits online students by providing them with
the flexibility they need and giving them more time to reflect on the course content. However,
all those characteristics that distinguish a face-to-face lesson are missing in asynchronous con-
texts (e.g., the opportunity of asking questions and the feeling of being part of a class). There-
fore, synchronous lessons are usually preferred while moving these face-to-face online.
   Tracking students’ behavior in both delivering modes is producing a huge amount of data
that can open new ways for better understanding how students learn and, therefore, analyze
data coming these contexts is becoming an unprecedented opportunity for the research and
educational research communities. A vast amount of studies on asynchronous online learning
has been carried out in literature [5, 6, 7, 8, 9, 10], mainly dealing with the analysis of click-
streams to predict whether a student will drop out of an online course [11, 12] or the final grade
a student will receive [13, 14]. Synchronous online learning is conversely an under-explored
domain and, given its growing adoption, there is an urgent need of analyzing how students
learn and how their behavior can be modelled under this educational modality as well [15].
   In this paper, we analyze how students interact with a synchronous online learning platform
in all the courses delivered by a university over a semester. For this purpose, we first collect
and pre-process student’s interactions in the form of entry-exit records from lesson rooms.
Then, we apply clustering techniques to model behavioral patterns shared between students
at faculty and course level. Finally, we identify and interpret the clusters, depicting their main
characteristics. More precisely, we will answer to the following research questions:
   1. How does the level of student’s participation in courses change over a semester?
   2. Which are the principal participation patterns at faculty and course level?
   3. Does the hour of the day a course is delivered influence the level of participation?
   The rest of this paper is organized as follows: Section 2 describes related work. Section 3
introduces the educational context and the dataset considered in this paper, while Section 4
provides a preliminary analysis of the students’ participation level, the participation strategy
modelling, and their analysis over the semester. 5, we represent the discussion, implications
and limitations of our work. Finally, Section 6 provides insights for future work.


2. Related Work
Our research lies at the intersection between studies on the analysis of synchronous learning
from an educational science perspective and those on educational data mining in the context
of online learning. To better contextualize our work, therefore, we discuss representative prior
work in these two fields and identify how we differ from them.
2.1. Synchronous Online Learning
Synchronous online learning is one of the most widely adopted methods by teachers due to
several reasons, defined as a real-time, instructor-led online learning experience where partic-
ipants are connected at the same time and communicate with each other [16]. For instance,
students can ask questions and get answers in real-time, as the lesson is live, and the instruc-
tor can assess and shape students’ understanding in real-time, adjusting his/her spoken words
accordingly. Moreover, students feel an increased sense of the instructor actually being there
and instructors can facilitate workshop-style classes and run breakout group activities, live
chats or office hours allow for real-time interaction (e.g., a conversation and synchronous ses-
sions provide a schedule to help students who struggle with task initiation to stay on track).
To better understand the benefits of synchronous online learning, for instance, Francescucci
et al. [17, 18] investigated the effects of a novel online synchronous course format, based on
virtual, interactive, real-time, instructor-led rooms, on students’ learning outcomes and their
level of engagement, compared to a fully face-to-face format in classroom. Similarly, Kohnke
and Moorhouse [19] analyzed whether students’ perceptions about communities of inquiry
constructs reflect their behaviours in the synchronous online learning settings and in what
ways the interrelationships between constructs differ in terms of students’ perceptions and
behaviours. The widespread adoption of synchronous online courses has become more and
more evident during the last year, as demonstrated by the recent studies conducted by [20, 21],
whose goal was to evaluate learner’s behavior in synchronous online continuing education lec-
tures during the COVID-19 pandemic. Some high schools found themselves in crisis to manage
online learning [22]. Indeed, it has become clearer that the education system is susceptible to
external dangers [23]. Feldman [24] addresses student assessment during this pandemic on how
districts can legislate unbiased and evenhanded grading policies based on these recommenda-
tions; (i) pandemic related anxiety will have negative effects on student academic performance,
(ii) academic performance of students might be affect by racial, economic and resource differ-
ences, and (iii) the larger parts of instructors were not effectively ready to deliver high-quality
instruction remotely.

2.2. Student Behavior Modelling
Educational data mining deals with the development of methods for exploring the increasingly
large-scale data obtained in educational settings [25, 26, 27, 28, 29]. Recent studies modelled
student’s behavioral patterns in terms of their participation in online initiatives (e.g., forums,
courses) and investigated the existence of relationships between students’ behavioral patterns
and their learning performance, assuming that different habits of online students will lead to
different levels of learning achievement in students [30]. Some studies focused their attention
on analyzing server logs, to construct knowledge on typical patterns of online learning behav-
iors, to further discover the unique advantages of data mining techniques to support dynamic
online instruction, and to build a predictive model for online learning [31]. In [32], data mining
techniques were applied to extracted LMS server logs to analyze and compare student online
learning behaviors between peer-moderated and teacher-moderated groups in an undergradu-
ate online collaborative project-based learning course in Taiwan. The authors in [33] compared
the performance of various clustering and classification algorithms applied on the same educa-
tional dataset. In [34], the authors provided an unexplored review of educational data mining
based on clustering with respect to teaching and learning processes. In [35, 36], a clustering
approach to partition students into different groups based on their learning behavior has been
presented. Furthermore, the authors presented the personalized e-learning system architec-
ture used to detect and adapt teaching contents according to the students’ level of knowledge.
Others, such as [37, 38], used the K-Means clustering algorithm to model behavioral patterns
of passing and failing students and others built effective indicators of students’ performance in
degree programmes. Our work differs from prior work in terms of educational context, given
that we tackle synchronous and not asynchronous online learning, while we rely on similar
techniques to model behavior in this emerging context (e.g., K-Means). We argue that behav-
ioral patterns are highly dependent from the educational context, so our study brings novel
observations and contributions to better understand the synchronous context.


3. Educational Context and Dataset
Our work considers data coming from synchronous learning delivered by an Italian university.
Specifically, the University of Cagliari has chosen to provide its lesson-related services through
Adobe Connect1 , an e-learning platform where students can access to virtual rooms by means
of computers, tablets and smartphones. For each student and each day of a given university
lesson, we analyzed their participation level based on the amount of lesson time attended by
the student (i.e., remained connected to virtual room). Before describing our approach, we
present the context, the university structure, and the data collected from Adobe Connect.
University Structure Description. More than 25,000 students have been covered by our
study, across the lessons of each degree programme. Six faculties have been involved, namely
Biology and Pharmacy, Engineering and Architecture, Medicine and Surgery, Science, Economic,
Law and Political Sciences, and Humanities. Figure 1 shows the number of degree programmes
per faculty, for a total of 89 degree courses. Each faculty nominates a person for coordinating
the lessons planning across programmes, booking then the virtual rooms in Adobe Connect.
Online Educational Infrastructure and Delivering. Each lesson has been delivered in an
Adobe Connect virtual room, which provides tools and services for teleconference, e-learning
sessions, and collaborative content creation and delivering. Each virtual room is a permanent
online space an instructor can use to deliver lessons to students. In the context of the university
we consider, more than 50 virtual rooms have been created, and each room was assigned to a
lesson of a specific course at a given time slot, similarly to classrooms assignments in a face-to-
face context. Then, each instructor is assigned to a virtual room on some fixed time slots across
weeks. Students enrolled in a specific course can login using their Adobe Connect account and
enter the virtual room associated to the desired lesson. As soon as the instructor starts the
lesson, students can take part in it. During the lesson, they can virtually raise the hand, write
messages in a chat, and ask the teacher to activate the video and audio in case they want to
answer to some questions or ask questions and then participate in the discussion. Figure 2
   1
       https://www.adobe.com/products/adobeconnect.html
Figure 1: Bachelor’s and Master’s degree programmes per faculty covered by our study (89 in total).




Figure 2: Schema of the online synchronous environment considered in this study.


provides a schematic representation of the Adobe Connect environment.
Data Collection Processes and Format. For each lesson, the date and time of entry and
exit of students from the virtual room, the content of the chat, and the reactions and raised
hands have been traced. However, for privacy reasons, this paper only deals with the analysis
of entry-exit data. This data is structured in folders, following a hierarchical structure aligned
with the faculty–degree-course structure of a university. Therefore, the main folder includes
six sub-folders, one per faculty. Inside each sub-folder, one Excel file with multiple sheets
pertaining to the entry-exit student’s records is stored for each course in that faculty. Each
sheet includes entry-exit records for a specific lesson of that course. Table 1 provides an
example of how a sheet is structured. We collected a total of 6,296 sheets across faculties and
degree programmes, with 684 sheets for the Faculty of Biology and Pharmacy, 1,320 sheets for
the Faculty of Engineering and Architecture, 947 sheets for the Faculty of Medicine, 760 sheets
for the Faculty of Science, 1,362 sheets for the Faculty of Economic, Law and Political Sciences,
and 1,223 sheets for the Faculty of Humanities (around 464.5 MB of data). This data covers an
entire semester, from March to June 2020.
                                                                           Example of a
 Attribute             Description                                         registered user
                                                                           (Guest)
                       The id of the transcription and it is unique
 transcript-id                                                             4445626093
                       for each line
                       The id related to that session. When a
                       teacher closes the classroom and ends the
 asset-id                                                                  4445465743
                       meeting, another one with a different id is
                       generated at the next login
 sco-id                The unique id of the virtual classroom              4211809190
                       The unique id of the registered user. For guest     3727616656
 principal-id
                       logins this field is empty                          (-)
                                                                           student@gmail.com
 login                 It is the username and for guests it is empty
                                                                           (-)
                       It is the name field, which is also present         2019_20/40/12345 John Smith
 session-name
                       for guests                                          (Guest Name)
                       It is name of the virtual classroom. Now it
 sco-name              has changed its name and it is called               CdL Letters - Room 3
                       "training object"
 date-created          Is the timestamp of entry into the classroom        2020-11-10 08:37:50
                                                                           2020-11-10 08:38:02
 date-end              It is the timestamp of leaving the classroom
                                                                           (-)
                       It is the registered username field and it is not   2019_20/40/12345 John Smith
 participant-name
                       present for guests                                  (-)
 answered-survey       If a survey is proposed, indicate who answered      0
Table 1
Example of data recorded for a lesson of a given course.


4. Student Behavior Characterization
In this section, guided by three research questions, we will analyze student participation pat-
terns throughout the semester, how student groups are related to the courses and if there is a
correlation between student participation and class schedule.
Participation Level. In this section, we aim to investigate the participation level of students
at faculty level in terms of the average number of students who attended courses in that fac-
ulty, to understand whether the trend of the participation level changes over the semester. As
mentioned in Section 3, Adobe Connect collects entry-exit logs in forms of CSV files. For each
CSV pertaining to a given lesson of a course, we counted the number of students present in
each lesson. Cases in which someone from the university staff logged in, empty fields and
duplicates were excluded (e.g., some students, due to connection problems, logged in several
times, for this reason a single access for each student was considered, not counting the other
attempts). This procedure was repeated for all the lessons pertaining to a given faculty, aver-
aging the number of students present in a lesson on a given day across all the lessons delivered
for that faculty on that day. Figure 3 shows the participation level over the semester for the six
graphs considered faculties. It should be noted that we cannot compare the extent to which
Figure 3: Average number of students per lesson for the six faculties considered in this study.


the student participate in the lessons across faculties, given that each faculty has a different
number of students. We are thus interested in the general participation trend in each faculty,
individually. For each plot, the time span in which the lessons have been held is represented
on the x-axis, ranging from mid-March to early June. The average of the number of students is
indicated in the y-axis. All plots are characterized by a strong initial interest, but over time this
interest tends to decrease. This decline is very present in three faculties in particular, namely
Medicine and Surgery, Economic, Law and Political Science, and Humanities. Science is one
of the few that tends to maintain a constant student average, but even here, there is a decline
as the months go by2 . Based on these patterns we understood that there is a need for an adap-
tive load based on the needs of the varying faculties. This kind of graphs have been generated
and analyzed mainly to give support to students and teachers. In particular, if we focus our
attention on trends, it is immediately clear that over time, fewer and fewer students connect
to the platform. This fact can help teachers to make the lessons more interactive and perhaps
increase the attention threshold, trying to limit as much as possible the drop in students over
the months. While these plots point out on the possibility of adapting the technological infras-
tructure based on the changing participation level over time, it does not tell us to what extent
these patterns vary for each course in a given faculty. Therefore, the following section analyzes
the participation patterns at course level.
Group Modelling. In this section, we are interested in investigating whether there exist core
participation patterns in each course and how these patterns vary across courses. To this end,
due to the amount of courses included in the dataset and to better shape our analysis, we
focused only on the courses of a specific faculty and a specific degree programme, namely

    2
     Please, note that the peaks down in the plots are associated to lessons that were planned but not given then.
This highlights the need for a more fine-grained analysis, we left as a future work.
Science and its Bachelor’s Degree in Computer Science. Specifically, we have analyzed the second
semester, a period between March and June 2020, including 9 courses (4 courses of the first
year, 4 courses of the second year and 1 course of the first year). Given the large number
of faculties and consequently of courses, we have selected a degree course with significant
patterns that can also be found in other courses. We leave pattern analysis in other courses
besides computer science courses to future work. For all the courses in this Bachelor’s Degree,
all the Excel files with the corresponding entry-exit records have been examined. For each
course and lesson in that course, we measure the amount of minutes a student was connected
to that lesson during that day, i.e. how long that particular student was connected to the
platform. Therefore, for each student in a given course, we computed a N-dimensional vector
(where N is the total number of lessons in that course) and each cell contains the amount of
minutes that student was connected into the platform for that lesson. Then, we computed the
pairwise distances across vectors and feed the pairwise-distance vectors (one per student) of
all the students in that course to a K-Means clustering algorithm. In order to identify the right
number of clusters, we applied the Elbow method, using the Silhouette score as a measure of
the quality of the clusters. Figure 4 shows the centroids of the clusters identified through the
applied methodology for each considered course. It can be observed that, for almost all the
courses, two main clusters of students have been identified and these two clusters are well-
balanced in terms of the number of students they include. One cluster refers to the students
who constantly follow the course over the semester, while the second cluster identifies students
whose level of participation decreases over time. This finding is particularly important as a first
point for shaping adaptive interventions to students who would need to be motivated along the
course. Interestingly, the course in the plot (d) was characterized by three main clusters. For
this course, there is a clear subset of students who do not engage with the course at all, from its
beginning. Observing this also highlights the need of multiple level of adaptive interventions3 .
Considering the participation patterns with respect to the hour of the day a course is delivered,
we observed that there is a relation between the slope of the curves for the cluster who lose
engagement and the time of the day, for the courses delivered on the same year of study. This
would be important for the future planning of the lessons schedule.


5. Implications and Limitations
There are some limitations on our work, but these do not compromise reliability of this pre-
liminary study. We only used the entry and exit data of the students’ connection, without
considering the intermediate intervals, so we do not manage cases where a student may have
been connected for a short time and have made many intermediate accesses, without perhaps
following the lesson for a certain period. Furthermore, due to privacy constraints, no addi-
tional information from the students was considered, e.g., all student interventions via chat or
    3
      It should be noted that the line associated to each cluster has sometimes a peak towards zero, e.g., in plot (a) in
Figure 4 for the lesson with ID 3. By carefully observing the data, it is possible to notice that for that lesson only two
students were present. For the peak with the lesson with ID 7, the students connected for a maximum of 30 minutes.
This could mean that the platform had problems and the connection was interrupted; another alternative would be
that the teacher assigned an exercise and so students no longer needed to stay online. Similar circumstances were
present for the other courses.
      (a) 1st Year                      (b) 2nd Year                      (c) 1st Year
          9.00 - 10.40 a.m. (Mon)           9.00 - 10.40 a.m. (Wed)           9.00 - 10.40 a.m. (Tue - Thu)




      (d) 1st Year                       (e) 1st Year                     (f) 3rd Year
          11.00 - 12.40 a.m. (Tue - Thu)     11.00 - 12.40 a.m. (Thu)         11.00 - 12.40 a.m. (Mon - Tue)




      (g) 2nd Year                     (h) 2nd Year                       (i) 2nd Year
          3.00 - 4.40 p.m. (Mon - Wed)     3.00 - 4.40 p.m. (Thu - Fri)       3.00 - 4.40 p.m. (Tue)
Figure 4: Centroids of the clusters identified for the nine considered courses.


system interactions (raised hand) were not included in this study. A limited period was taken
into consideration, given that the data refers to three months of lessons (i.e., a semester). This
means that, taking into account a longer period of time, different findings could emerge, so it is
our task to extend the work and analyze longer periods of time. The process has not considered
issues associated with connection problems. This can cause, as we have seen previously, some
problems on the analysis, in particular on the entry-exit records. It is not clear to us if users
who connected for short periods, less than 20 minutes, decided to leave the lesson or if they
had connection problems and were unable to reconnect shortly. Lastly, we used the K-Means
algorithm, but other techniques can be adopted to model the data, so it will be our future task
to try to model the data using other techniques and make comparisons.
6. Conclusions and Future Work
In this paper we analyze how students interact through the use of Adobe Connect, a syn-
chronous online learning platform, by means of data extracted from the platform logs. Then,
we modelled the students’ behavior, measuring the participation of students for each lesson
across an entire semester on all the courses and faculties of a university. Based on the results:

    • The workload associated to the technological infrastructure varies across the semester,
      with the first period subjected to higher workloads. This finding can be used to adapt
      the computational resources associated to the platform across time.

    • Half of the students often exhibited an strong decrease in attention along the semester,
      with the majority of them following regularly at beginning, then decreasing consider-
      ably. In some degree programs, this behavior is more evident.

    • The patterns exhibited in the analyzed data open up to adaptive interventions strategies,
      such as notifying the teacher in case there are students who are loosing engagement or
      who have stopped following, in order to understand the reasons and try to keep them
      engaged.

    • There is a relation between the participation level and the time of the day the courses
      are delivered, with a more rapid decrease in participation for courses delivered earlier in
      the day. This would be an important element to consider for the future planning of the
      lessons schedule.

   Our work opens up a wide range of future research directions. We plan to enrich the amount
of data analyzed, gradually inserting all the logs relating to the lessons that are done at the
moment and those done in the first half of 2020 (September-December). Furthermore, we will
extend our analysis to other approaches, taking into account chats, student reactions, raised
hands. It would be also interesting to make a comparison with students’ participation in face-
to-face courses in the pre-pandemic period and explore whether the behavioral patterns are
similar to those we showed in this paper. Finally, our findings can be of help for teachers to
understand the students’ participation level during their lessons, given that it is important that
a teacher is aware of how many students are present and how they follow the lessons.


Acknowledgments
Roberta Galici gratefully acknowledges the University of Cagliari for the financial support of
her PhD scholarship.


References
 [1] S. Hrastinski, Asynchronous and synchronous e-learning, Educause quarterly 31 (2008)
     51–55.
 [2] A. T. Peterson, P. N. Beymer, R. T. Putnam, Synchronous and asynchronous discussions:
     Effects on cooperation, belonging, and affect., Online Learning 22 (2018) 7–25.
 [3] G. Fenu, M. Marras, Leveraging continuous multi-modal authentication for access control
     in mobile cloud environments, in: New Trends in Image Analysis and Processing - ICIAP
     2017 - ICIAP International Workshops, volume 10590 of Lecture Notes in Computer Science,
     Springer, 2017, pp. 331–342.
 [4] S. Barra, M. Marras, G. Fenu, Continuous authentication on smartphone by means of
     periocular and virtual keystroke, in: Network and System Security - 12th International
     Conference, NSS 2018, volume 11058 of Lecture Notes in Computer Science, Springer, 2018,
     pp. 212–220.
 [5] F. Yang, F. W. Li, Study on student performance estimation, student progress analysis, and
     student potential prediction based on data mining, Computers & Education 123 (2018) 97–
     108.
 [6] M. Cantabella, R. Martínez-España, B. Ayuso, J. A. Yáñez, A. Muñoz, Analysis of stu-
     dent behavior in learning management systems through a big data framework, Future
     Generation Computer Systems 90 (2019) 262–272.
 [7] T. Lubicz-Nawrocka, K. Bunting, Student perceptions of teaching excellence: an analysis
     of student-led teaching award nomination data, Teaching in Higher Education 24 (2019)
     63–80.
 [8] Y. Wang, W. M. White, E. Andersen, Pathviewer: Visualizing pathways through student
     data, in: Proceedings of the 2017 CHI Conference on Human Factors in Computing Sys-
     tems, 2017, pp. 960–964.
 [9] P. Prinsloo, S. Slade, M. Khalil, Student data privacy in moocs: A sentiment analysis,
     Distance Education 40 (2019) 395–413.
[10] L. R. Amir, I. Tanti, D. A. Maharani, Y. S. Wimardhani, V. Julia, B. Sulijaya, R. Puspitawati,
     Student perspective of classroom and distance learning during covid-19 pandemic in the
     undergraduate dental study program universitas indonesia, BMC medical education 20
     (2020) 1–8.
[11] C. G. Brinton, S. Buccapatnam, M. Chiang, H. V. Poor, Mining mooc clickstreams: Video-
     watching behavior vs. in-video quiz performance, IEEE Transactions on Signal Processing
     64 (2016) 3677–3692.
[12] G. Kőrösi, R. Farkas, Mooc performance prediction by deep learning from raw click-
     stream data, in: International Conference on Advances in Computing and Data Sciences,
     Springer, 2020, pp. 474–485.
[13] J. L. Harvey, S. A. Kumar, A practical model for educators to predict student performance
     in k-12 education using machine learning, in: 2019 IEEE Symposium Series on Computa-
     tional Intelligence (SSCI), IEEE, 2019, pp. 3004–3011.
[14] R. I. Rashu, N. Haq, R. M. Rahman, Data mining approaches to predict final grade by over-
     coming class imbalance problem, in: 2014 17th International Conference on Computer
     and Information Technology (ICCIT), IEEE, 2014, pp. 14–19.
[15] Q. Cao, T. E. Griffin, X. Bai, The importance of synchronous interaction for student sat-
     isfaction with course web sites, Journal of Information Systems Education 20 (2009) 331.
[16] D. Leeder, H. Wharrad, T. BChir, Beyond institutional boundaries: reusable learning ob-
     jects for multi- professional education (2002).
[17] A. Francescucci, L. Rohani, Exclusively synchronous online (viri) learning: The impact
     on student performance and engagement outcomes, Journal of marketing Education 41
     (2019) 60–69.
[18] A. Francescucci, M. Foster, The viri (virtual, interactive, real-time, instructor-led) class-
     room: The impact of blended synchronous online courses on student performance, en-
     gagement, and satisfaction., Canadian Journal of Higher Education 43 (2013) 78–91.
[19] Ü. Çakıroğlu, S. Kılıç, Understanding community in synchronous online learning: do
     perceptions match behaviours?, Open Learning: The Journal of Open, Distance and e-
     Learning 35 (2020) 105–121.
[20] X. Yang, D. Li, X. Liu, J. Tan, Learner behaviors in synchronous online prosthodontic
     education during the 2020 covid-19 pandemic, The Journal of prosthetic dentistry (2020).
[21] T. C. Shoepe, J. F. McManus, S. E. August, N. L. Mattos, T. C. Vollucci, P. R. Sparks, Instruc-
     tor prompts and student engagement in synchronous online nutrition classes, American
     Journal of Distance Education 34 (2020) 194–210.
[22] O. B. Adedoyin, E. Soykan, Covid-19 pandemic and online learning: the challenges and
     opportunities, Interactive Learning Environments (2020) 1–13.
[23] A. Bozkurt, R. C. Sharma, Emergency remote teaching in a time of global crisis due to
     coronavirus pandemic, Asian Journal of Distance Education 15 (2020) i–vi.
[24] J. Feldman, To grade or not to grade, Educational Leadership 77 (2020) 43–46.
[25] O. Scheuer, B. M. McLaren, Educational Data Mining, Springer US, Boston, MA,
     2012, pp. 1075–1079. URL: https://doi.org/10.1007/978-1-4419-1428-6_618. doi:10.1007/
     978-1-4419-1428-6_618.
[26] D. Dessì, M. Dragoni, G. Fenu, M. Marras, D. R. Recupero, Evaluating neural word em-
     beddings created from online course reviews for sentiment analysis, in: Proceedings of
     the 34th ACM/SIGAPP Symposium on Applied Computing, SAC 2019, ACM, 2019, pp.
     2124–2127.
[27] L. Boratto, G. Fenu, M. Marras, Connecting user and item perspectives in popularity
     debiasing for collaborative recommendation, Inf. Process. Manag. 58 (2021) 102387.
[28] D. Dessì, G. Fenu, M. Marras, D. R. Recupero, Leveraging cognitive computing for multi-
     class classification of e-learning videos, in: The Semantic Web: ESWC 2017 Satellite
     Events - ESWC 2017 Satellite Events, Portorož, volume 10577 of Lecture Notes in Computer
     Science, Springer, 2017, pp. 21–25.
[29] G. Fenu, M. Marras, M. Meles, A learning analytics tool for usability assessment in moodle
     environments, Journal of e-Learning and Knowledge Society 13 (2017).
[30] J. Byun, D. Pennington, J. Cardenas, S. Dutta, J. Kirwan, Understanding student behaviors
     in online classroom: data scientific approach, in: 2014 IEEE International Congress on
     Big Data, IEEE, 2014, pp. 802–803.
[31] J.-L. Hung, K. Zhang, Revealing online learning behaviors and activity patterns and mak-
     ing predictions with data mining techniques in online teaching, MERLOT Journal of
     Online Learning and Teaching (2008).
[32] J.-L. Hung, S. M. Crooks, Examining online learning patterns with data mining techniques
     in peer-moderated and teacher-moderated courses, Journal of Educational Computing
     Research 40 (2009) 183–210.
[33] R. Ahuja, A. Jha, R. Maurya, R. Srivastava, Analysis of educational data mining, in:
     Harmony Search and Nature Inspired Optimization Algorithms, Springer, 2019, pp. 897–
     907.
[34] M. W. Rodrigues, S. Isotani, L. E. Zarate, Educational data mining: A review of evaluation
     process in the e-learning, Telematics and Informatics 35 (2018) 1701–1717.
[35] S. Kausar, X. Huahu, I. Hussain, Z. Wenhao, M. Zahid, Integration of data mining cluster-
     ing approach in the personalized e-learning system, IEEE Access 6 (2018) 72724–72734.
[36] D. Ding, J. Li, H. Wang, Z. Liang, Student behavior clustering method based on campus big
     data, in: 2017 13th International Conference on Computational Intelligence and Security
     (CIS), IEEE, 2017, pp. 500–503.
[37] M. Durairaj, C. Vijitha, Educational data mining for prediction of student performance
     using clustering algorithms, International Journal of Computer Science and Information
     Technologies 5 (2014) 5987–5991.
[38] R. Asif, A. Merceron, S. A. Ali, N. G. Haider, Analyzing undergraduate students’ perfor-
     mance using educational data mining, Computers & Education 113 (2017) 177–194.