<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Feedback Focus: A Tool for Evaluating and Reflecting on Instructor to Student Feedback Communication</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Derek Maki</string-name>
          <email>dmaki@mhc.ab.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sabine Graf</string-name>
          <email>sabineg@athabascau.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Athabasca University, School of Computing and Information Systems</institution>
          ,
          <addr-line>Alberta</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Medicine Hat College, School of Trades and Technology</institution>
          ,
          <addr-line>Alberta</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A critical challenge in education is enabling instructors to provide adaptive and constructive written feedback that supports life-long learning and enhances communication with learners. Effective feedback not only influences student engagement but also drives their learning progression. Currently, instructors provide feedback on assessments without tools to analyze or adapt their feedback practices in response to learner needs. Addressing this gap is essential to fostering a dynamic, adaptive, and supportive learning environment. This paper introduces Feedback Focus (FeeFo), a software tool designed to help educators adapt their feedback practices by offering actionable insights through advanced dashboards. FeeFo collects and analyzes written feedback, enabling instructors to identify trends and refine their communication strategies across courses, assessments, and individual learners. By empowering educators to visualize and track feedback effectiveness (e.g., through sentiment, emotions, and grades), FeeFo facilitates informed decision-making, fostering a cycle of continuous improvement in teaching practices. By supporting adaptive teaching methods, FeeFo enhances instructor-student interaction and fosters life-long learning for both educators and students. For instructors, FeeFo provides a pathway to continually refine their feedback practices, helping them develop the skills needed to craft more impactful and constructive feedback over time. For students, the tool encourages life-long learning by ensuring the feedback they receive is constructive in nature, fostering a sense of support and progress. By promoting meaningful, adaptive communication, FeeFo empowers educators to inspire persistence and growth in their students while also advancing their own professional development in delivering effective feedback.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;teaching analytics</kwd>
        <kwd>teacher feedback</kwd>
        <kwd>teacher dashboard</kwd>
        <kwd>feedback analysis 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Providing students with constructive written feedback from instructors is a pivotal element in the
educational process [5, 24], significantly impacting student engagement and their subsequent
learning progression [16]. Beyond immediate performance improvement, feedback serves as a vital
mechanism for fostering adaptive, life-long learning. Despite its importance, the current practice
within educational institutions largely overlooks the analysis of such feedback, missing out on
invaluable insights for instructors to improve the way they write such feedback. This lack of
reflection on feedback practices restricts the ability of instructors to adapt their teaching methods
effectively and undermines their capacity to foster constructive communication that supports
lifelong learning. Addressing this gap is crucial to enabling both students and instructors to thrive in
an educational environment. As instructors continue to provide feedback for assessments, there is a
shortfall in the feedback loop, as the communication between instructors and students is rarely
analyzed or assessed. As such, there is a need to assist educators in refining their feedback process
and reflecting on how their feedback have an impact on learners[18].</p>
      <p>To address this gap, the central focus of this research is on how to design an effective tool that
can collect, analyze, and visualize the communication that occurs within instructor feedback data,
based on the experiences and recommendations from literature. As a result, this paper introduces
the intelligent software tool Feedback Focus (FeeFo) which is suitable for use at all instructional
levels but is anticipated to be especially beneficial for secondary and post-secondary courses in
online and blended delivery. By helping educators adapt their communication strategies and
feedback practices through longitudinal analysis, FeeFo promotes a culture of continuous
professional development for instructors and sustained engagement for students. FeeFo allows
educators to import feedback and performance data from well-known learning management
systems (i.e., Moodle and Blackboard). To analyze the data, FeeFo uses quantitative and qualitative
(sentiment and emotion) analysis techniques and then visualizes the data through three main
dashboards, providing an overview of instructor derived feedback communication at the course,
assessment, and student level across longitudinal periods. FeeFo detects and quantifies specific
aspects such as emotional tone, sentiment patterns, and the volume of feedback. The tool integrates
this analysis with student performance data over time, providing dashboards to visualize the
relationship between instructor feedback and its effects on student learning outcomes.</p>
      <p>This tool empowers instructors to craft feedback that is constructive and adaptive, ensuring it
remains supportive rather than overly critical. By making instructors aware of how their feedback
may be perceived, where they dedicate most of their time in writing feedback, and how their
feedback affects learners, instructors can gain valuable insights for improving their feedback
practices. This tool not only assists educators in refining their communication through feedback
but also fosters a supportive learning environment where students are encouraged to persist and
thrive, embracing challenges as part of their life-long learning journey. In addition, such insights
can also provide information on problems within the course (e.g., an assessment that requires
much more feedback than others) and could point out opportunities for improvement in the course
designs.</p>
      <p>This research intersects in the fields of Learning Analytics, Intelligent Systems, Teaching
Analytics, and Learning Design, pushing beyond the traditional boundaries of learning analytics
that primarily focus on student data [9, 22]. Putting an emphasis on teaching analytics and the
effective use of machine learning techniques help provide instructors with insights into their
feedback practices, identifying where they may need to adapt their communication strategies to
ensure students are encouraged to continue their education and not become discouraged.</p>
      <p>The paper is structured as follows: Section 2 outlines related literature. Section 3 describes the
main features of the proposed tool with an emphasis on the functionality of the end user
dashboards, the primary analysis techniques that are used to provide insight into feedback, and
how those dashboards can be utilized. Section 4 describes the architecture of the software tool and
the technical details including which components are in place to make the software tool function.
Finally, in Section 5 we conclude the paper and discuss the directions for future research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>Assessment feedback data exists in most courses and is the written words that an instructor
writes to the student regarding the assessment that has been completed throughout the course
usually associated with a mark on the assessment. Many research articles consider such feedback
to be an important aspect of education and that such feedback can either positively or negatively
influence a student’s learning experience as well as their outcome [8, 12, 14, 24]. For example,
Henderson and colleagues found “students to prefer comments that were framed positively and
focused on good aspects of the students’ performance” [8], reinforcing that the tone of feedback is
important.</p>
      <p>Just as feedback from educators to students is important for learning [26], feedback for
educators is equally important for the self-improvement of educators and their teaching practices
[12, 21]. When educators receive feedback about their teaching practice they can use this
information for reflection and to adapt their feedback practices[4, 19]. Course evaluations are often
the only way that educators receive feedback regarding their teaching methods. However, the
feedback that is provided by students in course evaluations about their educator are often biased
[23], often the feedback is overly negative [4], and it typically does not go into detail on the way an
educator writes their feedback to students. In most cases educators strive to improve their teaching
practice, but without feedback on their written feedback to students it can be difficult to reflect on
and improve this important skill. Presenting educators with information on how they provide
feedback to students would give those educators insight into their process and areas for potential
improvement.</p>
      <p>When it comes to analyzing feedback data, most of the research in the area focuses on the
analysis of student generated feedback data about the course and/or the educator [7, 15, 23]. Data
mining techniques such as text mining and summarization have been applied to student feedback
to help quantify data in student responses [3]. Sentiment analysis has been used extensively to
quantify the responses students provide in feedback [3, 20, 25]. Shaik et al. highlighted that natural
language processing (NLP) in the form of chat bots could be used to gather and analyze student
feedback and point out chapters of a course that could benefit from revision[20]. Okoye et al. used
emotion detection on student surveys to predict whether or not they would recommend the
instructor’s course after they had completed the course [15]. Feature extraction has also been used
to evaluate student opinions in feedback at the course, program and university level, providing
important insights into teaching and curriculum [20].</p>
      <p>While there is less research on analyzing feedback from educators compared to feedback from
students, findings suggest that such feedback can have a significant impact on student performance
[5, 15]. Data mining and NLP techniques have been successfully used to provide valuable insights
into student-generated feedback [3, 15, 20]. However, using such techniques for educator-generated
feedback, for example, to understand whether the feedback educators provide is positive or
negative and how it impacts learners, is equally critical for enhancing teaching strategies [2, 13].
Nicoll, Douglas and Brinton developed a method that uses NLP to analyze feedback given to
students and correlate it with changes in student grades [18]. Their research found a high level of
correlation between the way that feedback was constructed and future student performance, the
most critical feature being the inclusion of the student’s name at the beginning of the assessment
feedback [18]. The researchers also mention that very little work has been done to provide
educators with tools to help evaluate and craft effective feedback for students [18]. In addition, Lin
et al. used NLP techniques on the feedback given by the instructor on the first assessment to
extract features in feedback text that lead to both increased and decreased performance in the
second assessment [10]. Furthermore, based on their findings, related research by Dawson et al.
highlights the need for a tool that can not only look at feedback for a single assessment, but also
collect the data over longer periods of time such as an entire program [5].</p>
      <p>Rubie noted that negative feedback should be delivered in a manner that engages and energizes
the recipient, otherwise, it may not be perceived as constructive or seen as an opportunity for
growth [17]. Similarly, in their review of the literature, Mercer and Gulseren identified research
demonstrating that constructive criticism, characterized by a considerate tone, can be effectively
used as a form of negative feedback to promote improvement [13]. However, they note that
negative feedback can also sometimes lead to unintended consequences [13]. While there are
instances where negative feedback can benefit learners, it can also have adverse effects that vary
from one learner to another [13, 17]. Further research by Câmpean et al. reinforces the fact that
positive feedback has a profound effect on student motivation and that educators overwhelmingly
agree that providing positive feedback is an extremely important aspect of education [2]. To
improve the way in which educators write feedback to learners they need to receive feedback on
the words they write to learners so they can reflect on how the feedback is perceived, improve
their feedback writing skills and as such adapt the way in which they provide feedback to learners.
Integrating insights about the nature of positive and negative feedback into analysis tools would
not only help educators reflect on their practices [2] but could guide the development of future
software that offers a holistic view of their feedback's influence. However, to the best of our
knowledge, no tool exists that analyzes and visualizes feedback data from educators to allow them
to learn more about their feedback practices, reflect on how they provide feedback to students and
better understand potential shortcomings in their course designs.</p>
    </sec>
    <sec id="sec-3">
      <title>3. TOOL FEATURES</title>
      <p>The next subsections describe in detail the three main dashboards of the FeeFo tool, as well as
how they can benefit educators and support them in reflecting on their feedback practices at the
course, assessment, and individual student levels. The tool has been designed for educators at any
level of education. However, we see the tool being most effective in an online or blended learning
context at the secondary and post-secondary levels.</p>
      <p>Each dashboard in FeeFo shows results and visualizations that were found to be useful and
recommended by literature. For the visualizations that present the sentiment and emotion of
feedback, the overall sentiment and emotion data has been produced by existing machine learning
models that have been trained and verified[1, 6, 11].</p>
      <sec id="sec-3-1">
        <title>3.1. Course Dashboard</title>
        <p>The course dashboard allows educators to look at all collected instructor-to-student feedback
data from the course level, which includes all assessments such as assignments, quizzes, exams, etc.
The course dashboard is geared toward the summarization of the collected data by course. Data can
be filtered by year, semester, and course as viewed at the top of Figure 1, allowing educators to
extract insight from their feedback data at specific timeframes and courses. Each filter option
allows the selection of one or more courses, years, or semesters.</p>
        <p>At the top of Figure 1, just below the filter, there are four cards that show a summary of the data
that is being inspected as follows:




“Total Feedback Items” card tallies all feedback items/comments written by the
instructor. The “More info” link reveals a data table with individual feedback items.
“Average Words in Feedback” card shows the mean word count per feedback item.
The “More info” link shows detailed word count analytics.
“Min/Max Words in Feedback” card indicates the range of word counts in feedback.
The “More info” link provides a detailed breakdown by course.
“Total Courses” card counts the courses in the dataset. The “More info” link leads to a
detailed course list.</p>
        <p>Within the dashboard six visualizations are provided to help the educator better understand the
processed data:






“Overall Course Emotion” shows the aggregation of all emotion in the selected
courses.
“Course Emotion” breaks down the emotion observed in feedback from each course
that is selected.
“Course Sentiment” shows the sentiment of feedback in each selected course.
“Course Grade Distribution” shows student marks in all selected courses grouped
into ranges.
“Average Grade in Course” shows every selected course and what the average grade
of all assessments was in each year.
“Average Words in Course Feedback” lets the educator see how many words they
typically write to each student on average in each selected course.</p>
        <p>The Course Dashboard is a powerful visualization instrument that can help educators identify
how their feedback is being perceived as well as how much feedback is written in each course.
Looking at the overall sentiment and emotion aids the instructor in being mindful of how positive
or negative their writing is, or at least how it may be perceived. Keeping track of the average
amount of words in feedback gives an idea of how much time is spent marking assessments in the
courses. Keeping track of the grades overall and over several time periods may prompt the
instructor to reflect on the attainment of course objectives or the difficulty of each course.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Assessment Dashboard</title>
        <p>The Assessment Dashboard allows educators to investigate more specific data related to their
feedback and student outcomes at the assessment level. Educators can again filter by course,
semester and year, but additionally they can also filter by assessment name. Each filter option
allows again the selection of one or more courses, years, semesters, or assessments.</p>
        <p>As seen in Figure 2, the assessment dashboard provides the following cards to summarize the
data that is active in the dashboard:



“Total Words in Assessment Feedback” card sums up the words in all feedback
items for the chosen assessments. In addition, it provides information on how many
feedback items are analyzed. The “More info” link leads to the raw feedback items for
review.
“Most Prevalent Emotion in Feedback” card highlights the dominant emotion in the
feedback from the selected assessments. The “More info” link leads to the emotion
analysis for each feedback item.
“Feedback Length Comparison” card evaluates if the feedback length for the selected
assessments is below, at, or above the average compared to all data. The “More info”
link shows how the length of feedback in each assessment compares to the average
length of feedback across all assessments.</p>
        <p>The following six charts are presented in the Assessment Dashboard (as shown in Figure 2),
focusing on providing an overview of the feedback and grade data with respect to the selected
assessments:






“Overall Assessment Emotion” shows the aggregation of emotion across all selected
assessments.
“Assessment Emotion” breaks down individually the amount of each emotion for the
selected assessments.
“Assessment Sentiment” showing the positivity and negativity detected within the
feedback of each selected assessment.
“Assessment Grade Distribution” visualizes how students performed on the selected
assessments.
“Average Grade on Assessment” shows how students performed on average in each
selected assessment per year.
“Average Words in Assessment Feedback” shows the number of words that have
been written on average for each selected assessment, allowing a quick comparison
between assessments.</p>
        <p>The visualizations at the assessment level let the educator investigate detailed data regarding
emotion and sentiment at the assessment level. Assessments that show overly negative emotions
and sentiment may be an area where the instructor wishes to investigate why the feedback exhibits
those issues. Further, allowing the instructor to zero in on assessment data allows them to
investigate specific assessments that students consistently perform poorly on or where educators
give more feedback than the usual amount found in other assessments.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Student Dashboard</title>
        <p>This dashboard puts the emphasis on student specific data, allowing educators to choose a
selection of one or multiple students that have feedback data available. In addition, the dashboard
enables the user to filter based on one or more years, semesters and courses. With this dashboard,
instructor can investigate the learning outcomes of their students and how they have interacted
with those students through feedback.</p>
        <p>As seen in Figure 3 the student dashboard provides the following cards to summarize the data in
the dashboard:




“Number of Students” shows a number of students selected and being analyzed in the
dashboard. The “More info” link leads to the raw feedback entries for those students.
“Number of Courses” card displays the distinct number of courses in the dashboard
for selected students. The “More info” link displays a more detailed list of the courses.
“Total Words in Feedback” card shows the total number of words in all feedback
items for selected students. “More info” link shows the educator a detailed summary of
each student and the number of words in each of their feedback items.
“Average Sentiment Score” card shows the average sentiment score for all selected
students and their associated feedback items. More concretely, the percentage score for
positive, neutral and negative sentiment in the feedback items is displayed. The “More
info” link goes to a summary page with sentiment for each feedback item grouped by
student and assessment.</p>
        <p>The student dashboard also includes the following six charts:


“Overall Student Emotion” shows the aggregation of emotion across the feedback
items of all selected students.</p>
        <p>“Student Emotion” breaks down the individual emotions for the selected students.
“Student Sentiment” shows the positivity and negativity detected within the feedback
for each individual student who has been selected.
“Student Grade Distribution” visualizes how selected students performed on all
assessments.
“Average Grade by Student” shows the average grade on how selected students
performed overall on all assessments per study year.
“Total Words in Feedback by Student” shows the number of words that have been
written for each selected student.</p>
        <p>The Student Dashboard provides a granular view of the feedback and performance of individual
students. This feature enables educators to monitor how much feedback each student receives, how
they are doing in their assessments and track the sentiment and emotional tone of feedback
comments.</p>
        <p>Analyzing the feedback per student helps instructors identify patterns in how students of
varying performance levels are receiving and perceiving feedback. Additionally, tracking the
grades and progress of each student can assist in identifying students who may need extra support
or those who consistently excel, thereby helping to tailor teaching strategies to the needs of
individual learners.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Architecture</title>
      <p>The overall architecture of FeeFo is shown in Figure 4. FeeFo is built using C# in the .NET
ecosystem utilizing a single page application framework called Blazor and is designed to run in the
browser as a progressive web application. As a progressive application, once the user browses to
the tool’s URL, they are presented with an optional install button in the address bar for offline use.
Whether used online or offline, no user registration is required because the application is designed
to run in a stateless manner on the client’s device. The application continues to work even when
the user does not have an internet connection. The only limitation is that new sentiment and
emotion analyses cannot be conducted but previously analyzed data can still be loaded and
visualized.</p>
      <p>Designing the tool as stand-alone tool instead of a plugin within a learning management system
(LMS) has the advantage that instructors can immediately use it once they have data they want to
analyze, rather than having to go through lengthy institutional approval processes to get a plugin
integrated into the institutional LMS.</p>
      <p>To streamline the data transfer between FeeFo and the LMS, FeeFo offers an easy import feature.
Essential data such as student identifier, names, grades, and written feedback can be exported from
an LMS (e.g., Moodle or Blackboard) using the "Work Offline" feature. This data can then be
seamlessly imported into a pre-configured course assessment in FeeFo. In addition, a manual data
entry feature is available to manually enter or revise relevant data.</p>
      <p>The application is designed with a focus on security and privacy since sensitive data (i.e.,
student names, student identifier, feedback data, and grades) are stored. As such, the application
stores all sensitive information entirely on the user’s device instead of hosting the data on an
external database server. The only data that leaves the device is the written feedback which is sent
in anonymous form to an external API for sentiment and emotion analysis.</p>
      <p>To run the machine learning algorithms for sentiment and emotion analysis, the external API
connects to machine learning models hosted on hugging face [1, 6, 11]. This allows the user to
browse the application and immediately gain access to the machine learning models without
having to register for any services.</p>
      <p>FeeFo uses a Sqlite database that runs completely in the browser and is set up automatically
when the application is accessed. To connect to the database, an object relational mapper (ORM)
called Entity Framework (EF) is in place. Using EF allows the entire codebase to rely on objects that
can be easily extended or modified by developers who wish to clone and modify the source code.</p>
      <p>To facilitate an app that runs locally FeeFo utilizes the WebAssembly environment keeping the
entire application inside modern browsers such as Chrome and Firefox. The underlying Sqlite
database is a local file database that remains in the browser cache keeping all data stored locally.</p>
      <p>Since all data are initially stored only in the browser cache, FeeFo has a function that allows
users to store data locally as backup file and then restore them again in case the browser cache was
emptied. Making the user save the data in this way ensures sensitive student data remains private
and is never uploaded to a third party for storage.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This paper introduces FeeFo, a teaching analytics tool for collecting, analyzing and visualizing
instructor-to-student feedback data. FeeFo supports instructors in crafting feedback that is both
constructive and adaptive, empowering them to foster more effective communication with learners.
The analytics tool presents the instructor with three main dashboards that they can use to
investigate feedback data and student grades at course level, assessment level and student level.
Using data aggregation and machine learning techniques, the dashboards show different types of
information such as emotion and sentiment detected in the feedback sent to students, student
overall grades, average student grades per course, assessment or student, and average word count
of feedback items.</p>
      <p>Traditionally, instructors receive feedback on their performance primarily through student
evaluations conducted at the end of a course. However, these evaluations rarely provide insights
into the quality or impact of the feedback instructors give to students, limiting opportunities to
improve and adapt their teaching practices. FeeFo addresses this gap by enabling instructors to
analyze their feedback practices using visualizations that encourage reflection and
selfimprovement. As such, FeeFo enables educators to adapt their teaching practices to provide more
constructive feedback that better supports students. While models trained on data from social
media come with limitations for sentiment and emotion analysis, such as challenges in interpreting
the formal and nuanced language of an educational context, based on our experimentation they
still offer a good foundation for sentiment and emotion analysis.</p>
      <p>To ensure that FeeFo becomes a valuable tool for educators, future research aims to gather
supporting empirical data by evaluating the tool through educator usage and feedback on its
usefulness and usability. This iterative feedback process ensures that FeeFo continues to evolve to
meet the dynamic needs of educators and learners. Additional research considerations include
integrating new and emerging machine learning models such as Large Language Models (LLMs)
into the application for further analysis of feedback data. Furthermore, while the current version of
FeeFo focuses on areas like positivity, negativity, and emotion within feedback, future extensions
can explore additional aspects of feedback, such as its focus on specific tasks, processes, or
cognitive strategies, as well as how feedback is structured and delivered to enhance its
effectiveness.</p>
      <p>FeeFo is available at https://feefo.ca, and the source code is available for download and
modification at https://github.com/FeedbackFocus/FeeFo. FeeFo is released with an MIT License,
which means that contributors can modify the source code as they see fit. By allowing contributors
to integrate more powerful models or add new features, FeeFo encourages a collaborative approach
to enhancing teaching practices.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>The authors acknowledge the support of the National Science and Engineering Research
Council of Canada (NSERC) [RGPIN-2020-05837] as well as the Medicine Hat College.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT-4o to assist with grammar and
spelling checks. The author(s) reviewed and edited the content as needed and take(s) full
responsibility for the publication’s content.
[5] Dawson, S., Pardo, A., Salehian Kia, F. and Panadero, E. 2023. An Integrated Model of
Feedback and Assessment: From fine grained to holistic programmatic review. LAK23: 13th
International Learning Analytics and Knowledge Conference (New York, NY, USA, Mar. 2023),
579–584.
[6] Emotion English DistilRoBERTa-base: 2022.
https://huggingface.co/j-hartmann/emotionenglish-distilroberta-base/.
[7] Hadad, Y., Keren, B. and Naveh, G. 2020. The relative importance of teaching evaluation
criteria from the points of view of students and faculty. Assessment &amp; Evaluation in Higher
Education. 45, 3 (Apr. 2020), 447–459. DOI:https://doi.org/10.1080/02602938.2019.1665623.
[8] Henderson, M., Ryan, T., Boud, D., Dawson, P., Phillips, M., Molloy, E. and Mahoney, P. 2021.</p>
      <p>The usefulness of feedback. Active Learning in Higher Education. 22, 3 (Nov. 2021), 229–243.</p>
      <p>DOI:https://doi.org/10.1177/1469787419872393.
[9] Khuzairi, N.M.S., Sidhu, M.S. and Cob, Z.C. 2020. Learning Analytics and Teaching Analytics:
The Similarities and Differences. International Journal of Humanities, Management and Social
Science (IJ-HuMaSS). 3, 2 (Dec. 2020), 52–58.
DOI:https://doi.org/10.36079/lamintang.ijhumass-0302.135.
[10] Lin, J., Dai, W., Lim, L.-A., Tsai, Y.-S., Mello, R.F., Khosravi, H., Gasevic, D. and Chen, G. 2023.</p>
      <p>Learner-centred Analytics of Feedback Content in Higher Education. LAK23: 13th
International Learning Analytics and Knowledge Conference (New York, NY, USA, Mar. 2023),
100–110.
[11] Loureiro, D., Barbieri, F., Neves, L., Espinosa Anke, L. and Camacho-collados, J. 2022.</p>
      <p>TimeLMs: Diachronic Language Models from Twitter. Proceedings of the 60th Annual Meeting
of the Association for Computational Linguistics: System Demonstrations (Dublin, Ireland, May
2022), 251–260.
[12] Mandouit, L. and Hattie, J. 2023. Revisiting “The Power of Feedback” from the perspective of
the learner. Learning and Instruction. 84, (Apr. 2023), 101718.</p>
      <p>
        DOI:https://doi.org/10.1016/j.learninstruc.2022.101718.
[13] Mer
        <xref ref-type="bibr" rid="ref2">cer, M. and Gulseren, D.B. 2024</xref>
        . When negative feedback harms: a systematic review of
the unintended consequences of negative feedback on psychological, attitudinal, and
behavioral responses. Studies in Higher Education. 49, 4 (Apr. 2024), 654–669.
      </p>
      <p>DOI:https://doi.org/10.1080/03075079.2023.2248490.
[14] Morris, R., Perry, T. and Wardle, L. 2021. Formative assessment and feedback for learning in
higher education: A systematic review. Review of Education. 9, 3 (2021), e3292.</p>
      <p>DOI:https://doi.org/10.1002/rev3.3292.
[15] Okoye, K., Arrona-Palacios, A., Camacho-Zuñiga, C., Achem, J.A.G., Escamilla, J. and
Hosseini, S. 2022. Towards teaching analytics: a contextual model for analysis of students’
evaluation of teaching through text mining and machine learning classification. Education
and Information Technologies. 27, 3 (Apr. 2022), 3891–3933.</p>
      <p>DOI:https://doi.org/10.1007/s10639-021-10751-5.
[16] Rabia Aslam and Najmonnisa Khan 2020. Constructive Feedback and Students’ Academic
Achievement: A Theoretical Framework | New Horizons. New Horizons. 14, 28 (Nov. 2020).</p>
      <p>
        DOI:https://doi.org/10.2.9270/NH.14.2(20).10.
[17] Rubie,
        <xref ref-type="bibr" rid="ref2">Candice 2024</xref>
        . Investigating a Growth Frame for Delivering Effective Failure Feedback.
      </p>
      <p>University of Waterloo.
[18] Serena Nicoll, Kerrie Douglas, and Christopher Brinton 2022. Giving Feedback on Feedback:
An Assessment of Grader Feedback Construction on Student Performance. LAK22: 12th
International Learning Analytics and Knowledge Conference. (2022), 239–249.</p>
      <p>DOI:https://doi.org/10.1145/3506860.3506897.
[19] Sergis, S., Sampson, D.G., Rodríguez-Triana, M.J., Gillet, D., Pelliccione, L. and de Jong, T.
2019. Using educational data from teaching and learning to inform teachers’ reflective
educational design in inquiry-based STEM education. Computers in human behavior. 92,
(2019), 724–738.
[20] Shaik, T., Tao, X., Dann, C., Quadrelli, C., Li, Y. and O’Neill, S. 2022. Educational Decision
Support System Adopting Sentiment Analysis on Student Feedback. 2022 IEEE/WIC/ACM
International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)
(Nov. 2022), 377–383.
[21] Shaik, T., Tao, X., Li, Y., Dann, C., McDonald, J., Redmond, P. and Galligan, L. 2022. A Review
of the Trends and Challenges in Adopting Natural Language Processing Methods for
Education Feedback Analysis. IEEE Access. 10, (2022), 56720–56739.</p>
      <p>DOI:https://doi.org/10.1109/ACCESS.2022.3177752.
[22] Siemens, G. 2013. Learning Analytics: The Emergence of a Discipline. American Behavioral</p>
      <p>Scientist. 57, 10 (Oct. 2013), 1380–1400. DOI:https://doi.org/10.1177/0002764213498851.
[23] Stroebe, W. 2020. Student Evaluations of Teaching Encourages Poor Teaching and
Contributes to Grade Inflation: A Theoretical and Empirical Analysis.Basic and Applied Social
Psychology. 42, 4 (Jul. 2020), 276–294. DOI:https://doi.org/10.1080/01973533.2020.1756817.
[24] To, J., Tan, K. and Lim, M. 2023. From error-focused to learner-centred feedback practices:
Unpacking the development of teacher feedback literacy. Teaching and Teacher Education.
131, (Sep. 2023), 104185. DOI:https://doi.org/10.1016/j.tate.2023.104185.
[25] Toçoğlu, M.A. and Onan, A. 2021. Sentiment Analysis on Students’ Evaluation of Higher
Educational Institutions. Intelligent and Fuzzy Techniques: Smart and Innovative Solutions
(Cham, 2021), 1693–1700.
[26] Xu, W., Meng, J., Raja, S.K.S., Priya, M.P. and Kiruthiga Devi, M. 2023. Artificial intelligence in
constructing personalized and accurate feedback systems for students. International Journal of
Modeling, Simulation, and Scientific Computing. 14, 01 (Feb. 2023), 2341001.
DOI:https://doi.org/10.1142/S1793962323410015.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Barbieri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Espinosa</surname>
            <given-names>Anke</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            and
            <surname>Camacho-Collados</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.</surname>
          </string-name>
          <year>2022</year>
          .
          <article-title>XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond</article-title>
          .
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference (Marseille</source>
          , France, Jun.
          <year>2022</year>
          ),
          <fpage>258</fpage>
          -
          <lpage>266</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Câmpean</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bocoș</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rad</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crișan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maier</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tăușan-Crișan</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Triff</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Triff</surname>
          </string-name>
          , D.-G.,
          <string-name>
            <surname>Mara</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mara</surname>
            ,
            <given-names>E.-L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Răduț-Taciu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Todor</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baciu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neacșu</surname>
          </string-name>
          , M.-G.,
          <string-name>
            <surname>Dumitru</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Colareza</surname>
            ,
            <given-names>C.C.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Roman</surname>
            ,
            <given-names>C.E.</given-names>
          </string-name>
          <year>2024</year>
          .
          <article-title>Examining Teachers' Perception on the Impact of Positive Feedback on School Students</article-title>
          .
          <source>Education Sciences. 14</source>
          ,
          <issue>3</issue>
          (Feb.
          <year>2024</year>
          ),
          <fpage>257</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>DOI:https://doi.org/10.3390/educsci14030257.</mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Claessen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2022</year>
          .
          <article-title>Text Mining to Improve Education: An Evaluation of Text Mining in the Student Feedback Process</article-title>
          . University of Twente.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Cornes</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Torre</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fulton</surname>
            ,
            <given-names>T.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oza</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Teherani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>H.C.</given-names>
          </string-name>
          <year>2023</year>
          .
          <article-title>When students' words hurt: 12 tips for helping faculty receive and respond constructively to student evaluations of teaching</article-title>
          .
          <source>Medical Education Online</source>
          .
          <volume>28</volume>
          ,
          <issue>1</issue>
          (Dec.
          <year>2023</year>
          ),
          <fpage>2154768</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          DOI:https://doi.org/10.1080/10872981.
          <year>2022</year>
          .
          <volume>2154768</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>