<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The impact of ChatGPT on student performance in higher education</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nimród Mike</string-name>
          <email>nimrod.mike@uni-corvinus.hu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Krisztina Karsai</string-name>
          <email>karsai.krisztina@szte.hu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gábor Orbán</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexandra Bubelényi</string-name>
          <email>alexandra.bubelenyi@stud.uni-corvinus.hu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Csaba</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Norbert Nagy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gábor Polyák</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ChatGPT</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Higher Education</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Large Language Model (LLM)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Generative AI (GAI).</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Corvinus University of Budapest</institution>
          ,
          <addr-line>Fővám tér 8, 1093, Budapest</addr-line>
          ,
          <country country="HU">Hungary</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Eötvös Lóránd University</institution>
          ,
          <addr-line>Múzeum krt. 6-8, 1088 Budapest</addr-line>
          ,
          <country country="HU">Hungary</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Szeged</institution>
          ,
          <addr-line>Dugonics tér 13, 6720, Szeged</addr-line>
          ,
          <country country="HU">Hungary</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This study investigates ChatGPT's impact on effectiveness, efficiency, and problem-solving among higher education students in law, business informatics, and media and communication. Involving 304 students divided into experimental (using ChatGPT) and control groups for an open-book test, the research aimed to assess efficiency benefits. Contrary to expectations, ChatGPT did not improve performance across disciplines. However, business informatics students completed tests faster, suggesting a nuanced effect on efficiency.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Artificial Intelligence (AI) has significantly influenced education, with tools like ChatGPT
enhancing student efficiency, problem-solving, and understanding in higher education [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. While AI's educational potential is widely recognized, its specific effects on student
performance remain underexplored [9], [14]. Existing studies highlight AI’s role in
personalized learning and effective teaching strategies [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [34], but there is a lack of
empirical research comparing ChatGPT to traditional search tools like Google across various
disciplines [10]. This study aims to fill this gap by examining ChatGPT's impact on student
efficiency and problem-solving in higher education. Previous research, such as studies in
Ghana, generally focused on AI’s educational benefits [14].
      </p>
      <p>
        We expand on this by exploring ChatGPT's nuanced effects across different academic
disciplines, which has been underrepresented in the literature [22], [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], [21]. Our goal is to
provide insights that are relevant for both academic and policy-making contexts, helping
institutions leverage AI in education frameworks. We hypothesize that ChatGPT will
improve task efficiency but not necessarily enhance response accuracy compared to
traditional methods [16], [28]. This is based on the idea that AI tools streamline information
access and research processes, enhancing efficiency [26].
      </p>
      <p>
        Our study aims to inform educational policies on integrating AI tools like ChatGPT to
enhance teaching and learning [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Using a mixed-methods approach, we reviewed
literature and conducted an experiment with 304 Hungarian students, comparing ChatGPT
use in an open-book test to traditional methods. This research not only evaluates ChatGPT's
impact on academic performance but also explores broader implications for digital
governance and policy, reflecting growing interest in AI’s societal role [24]. Our findings add
empirical data on ChatGPT’s efficiency and problem-solving benefits, advocating for a
tailored approach to AI integration in education [13], [35]. By detailing ChatGPT's effects
across disciplines, we provide valuable insights for educators and policymakers interested
in the future of AI in education [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], [18].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>We employ a mixed-methods approach, combining qualitative and quantitative data, to
assess ChatGPT's impact on student performance. A thorough literature review informed
our research plan, emphasizing the value of diverse data collection techniques in capturing
the multifaceted effects of AI tools in education [15], [25].</p>
      <sec id="sec-2-1">
        <title>2.1. Literature review process</title>
        <p>To identify seminal publications within AI-related educational literature, we used
mathematical and statistical tools to identify key studies. We focused on articles with a high
number of authors and citations to ensure a comprehensive and diverse perspective. This
approach was chosen to capture a wide range of viewpoints and a strong academic impact.
We selected papers using the following formula:</p>
        <p>&gt;  3 + 1.5 ∗  (1)
where  is the observed value,  3 is the third quartile, and  is the interquartile range.
This statistical method helped us identify upward outliers, suggesting complexity and broad
acceptance in the scholarly community. We then created a directed graph to visualize the
interconnectedness of over sixteen hundred publications. This analysis revealed seven
subgraphs, with a dominant interconnected subgraph indicating a shared knowledge base.
Key publications within this subgraph were identified through their incoming edge degree
and PageRank, highlighting the most influential knowledge hubs.</p>
        <p>Our analysis provided a comprehensive overview of current research on ChatGPT in
higher education. We noted a significant study conducted in Ghana involving a flipped
classroom setup with 125 students, which demonstrated ChatGPT's positive impact on
critical, creative, and reflective thinking [14]. Recent research extensively explores the
integration and impact of generative AI (GAI) tools like ChatGPT across various educational
and professional contexts.</p>
        <p>Recent works emphasize the importance of AI literacy and practical learning, suggesting
research directions to prepare students for a society increasingly powered by GAI [9]. The
technological evolution of digital writing and summarization frameworks has also been a
focus, illustrating how ChatGPT can facilitate more advanced educational applications [22],
[21].</p>
        <p>
          Additionally, the transformative potential of ChatGPT and its implications for redefining
academic "originality" are explored in studies that discuss how ChatGPT challenges
traditional notions of academic integrity, urging educational institutions to adapt their
policies accordingly [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], [23]. Other papers discuss necessary adaptations to teaching and
assessment practices considering GAI, highlighting how educational frameworks need to
evolve to accommodate these technologies [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], [35].
        </p>
        <p>
          In addition to educational impacts, several studies address the broader ethical and
societal implications of AI tools like ChatGPT. Some works examine the benefits and ethical
concerns associated with using ChatGPT for scientific writing, noting the need for ethical
guidelines in its application [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. There is also analysis of varying public sentiments towards
GAI, indicating a diverse range of opinions on its integration into society [24]. Foundational
work and ethical considerations provide critical context for understanding the complex
implications of deploying AI in educational and professional settings [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], [13].
        </p>
        <p>Overall, the literature review emphasized the need for nuanced research into ChatGPT's
impact across different educational contexts, aligning with our study's objectives to explore
these effects in Hungarian universities. Recent research underscores the extensive
application of ChatGPT in enhancing educational practices and highlights the importance of
addressing the ethical and societal impacts of integrating AI tools in education.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Exploratory interviews</title>
        <p>To contribute to our literature review, we conducted semi-structured interviews with four
expert instructors from law, business informatics, and media studies. These interviews
provided qualitative insights into the experiences and perceptions of educators regarding
AI tools like ChatGPT.</p>
        <p>Primary evidence from these interviews includes direct quotes and specific observations
about the impact and challenges of integrating AI tools in higher education. For example,
one instructor noted, "ChatGPT has significantly enhanced the speed at which students can
gather initial research, but it lacks depth in more specialized areas," reflecting a common
sentiment among participants. Supporting information includes additional context
provided by the instructors, such as their perspectives on how AI tools are transforming
teaching methodologies and academic integrity.</p>
        <p>The interview protocol involved a set of guiding questions designed to explore various
aspects of AI integration, including how do they perceive the role of ChatGPT in enhancing
student learning and research capabilities; what are the main benefits and challenges they
have encountered in using ChatGPT in their teaching practices, if they do so; and how do
they address concerns regarding academic integrity and plagiarism in the context of AI
tools.</p>
        <p>Instructors expressed enthusiasm about ChatGPT's potential for idea generation but
highlighted concerns about its limitations in specialized research areas. Discussions
centered around the challenges of detecting AI-generated text and the implications for
academic integrity, suggesting that university policies, rather than new legislation, should
address these issues. This viewpoint emphasizes the need for institutional guidelines
tailored to AI tools' unique challenges.</p>
        <p>The interviews also revealed a consensus that academic tasks should evolve to
incorporate AI tools, shifting the focus from memorization to critical thinking and
problemsolving. Instructors proposed revising assignments to explicitly include the use of ChatGPT,
thus teaching students how to effectively utilize AI tools. They emphasized that while AI
might impact various professions, jobs requiring personal interaction would remain largely
unaffected [24]. These insights suggest a direction for developing new educational
strategies that integrate AI tools in a way that is both effective and ethically sound. As a
conclusion, the interviews gave useful insights into how AI tools like ChatGPT are being used
in education. They demonstrated the need to adjust our educational practices to include
these tools in a way that keeps academic integrity intact and continues to develop critical
thinking skills among students.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Pre-study preparations</title>
        <p>We recruited undergraduate students from various Hungarian universities, including those
studying law, business informatics, and media studies. Initially, 415 students enrolled in the
study, providing a diverse sample that represents key academic disciplines relevant to our
research. Participants were recruited through university email lists, ensuring a wide
outreach.
21.5 minutes.</p>
        <p>To establish a baseline for the research, participants completed an assessment designed
to evaluate their initial knowledge and attitudes towards ChatGPT. The assessment
included tasks to measure critical thinking and problem-solving strategies, such as
analyzing case studies, solving logical puzzles, and summarizing complex texts. Participants
had 30 minutes to complete the assessment, with the average completion time recorded at</p>
        <p>Based on their performance in the pre-test, participants were divided into an
experimental group (which utilized ChatGPT) and a control group (which relied on
traditional methods). The assignment process involved balancing scores using a weighted
formula that favored complex tasks, calculated as follows:
 
=

∑

=1
⬚ 


∗</p>
        <p>1
 
(2)</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Test designs</title>
        <p>Developed in close collaboration with domain experts, the tests for law studies, business
informatics, and media and communication Studies were structured to critically evaluate
the efficacy of AI-assisted learning tools like ChatGPT within higher education. Each
discipline's test consisted of 30 questions, employing a uniform format that included
true/false, multiple-choice, fill-in-the-blank, and matching questions. This approach was
chosen to test a broad spectrum of knowledge and application skills across different
educational domains, with the curriculum serving as the basis for question selection in each
specific area.</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Proprietary data collection tool</title>
        <p>To gather comprehensive data, we developed ExamEye, a specialized browser extension
that captured student interactions with ChatGPT and traditional search engines during
controlled tests. ExamEye prioritized ethical standards and privacy, activating only within
the testing platform and automatically ceasing recording upon test completion.</p>
        <p>ExamEye provided a rich dataset, tracking participants' digital activity throughout the
test environment. It recorded browsing activity, source type, and engagement with
ChatGPT, including prompt crafting and response evaluation. This allowed us to distinguish
between the use of ChatGPT and traditional research methods and analyze internet usage
patterns. The tool's design ensured that only relevant data was collected, minimizing any
potential privacy concerns.</p>
        <p>Participants provided informed consent and were fully briefed on the use of ExamEye.
Privacy safeguards included anonymizing data and restricting monitoring to the test
environment only. The tool deactivated immediately upon test completion, ensuring that
data collection adhered to ethical standards and protected participant privacy. Participants
were informed about the data being collected and how it would be used, ensuring
transparency and compliance with ethical guidelines.</p>
      </sec>
      <sec id="sec-2-6">
        <title>2.6. Data collection and analysis</title>
        <p>Data was collected using ExamEye, which monitored student interactions with AI tools
during the tests. The data included usage patterns, time spent on tasks, and the nature of
the interactions with ChatGPT. Statistical analysis involved comparing test completion
times and accuracy between the experimental and control groups using t-tests and variance
analysis to assess differences.</p>
        <p>We employed independent sample t-tests to compare the means of test completion times
and accuracy between the two groups, assessing whether the differences were statistically
significant. Where variances were unequal, Welch’s t-test was used to ensure robust results.
Additionally, we performed sensitivity analyses to account for any variations in baseline
performance that could influence the outcomes. This comprehensive approach ensured that
our analysis was rigorous and reliable, providing clear insights into the impact of ChatGPT
on student performance.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <p>The research hypothesis posits that the average score of students in the experimental group
will be the same as that of the control group, with significantly better test times per stratum,
suggesting that the use of ChatGPT will primarily affect efficiency. To confirm or refute these
hypotheses, we first need to examine the significance of the differences between the
expected values of the test score (percentage) and test time (test_interval_s) variables for
each stratum of the experimental and control groups separately.</p>
      <p>The first step in hypothesis testing is to generate appropriate null and alternative
hypotheses based on our assumptions. We pair our null hypothesis that the experimental
group's outcome is identical to that of the control group on a stratified basis with partially
overlapping alternative hypotheses, because the one-tailed alternative tests have a higher
power of test than the two-tailed alternative test at the same level of significance, so that by
comparing the p-value of each test we can draw a more accurate conclusion.</p>
      <p>In order to test our hypothesis, we performed two independent sample t-tests to analyze
the disparity of means. Prior to conducting these hypothesis tests, we ensured that all
fundamental assumptions of the t-test were satisfied. This involved assessing the expected
normality within small sample sizes (although the t-test remains robust to non-normality
with sufficiently large sample sizes due to the central limit theorem) and confirming equal
variances between the control and experimental groups within each stratum. In cases
where significant differences in variances were observed, we opted for Welch’s t-test over
Student’s t-test, as it accommodates both unequal variances and sample sizes. Once we
confirmed that all test criteria were met, we proceeded with conducting the t-tests for each
individual stratum.</p>
      <p>Based on the findings, neither the informatics test nor the law test exhibited significant
differences between the control and experimental groups in terms of both magnitude and
direction. Thus, the observed variances could be attributed to random sampling, indicating
that the discrepancies observed cannot be generalized to the broader participants.
However, concerning the communication test, while the two-sided test indicated the
difference as insignificant, the one-sided test, which is more sensitive to directional
differences, deemed the gap significant with a p-value of 2.612%.</p>
      <p>In the table provided, the test results regarding the mean test time difference are
outlined. For the law test, despite observing a decrease in test time, the t-test deemed the
difference as statistically insignificant in both magnitude and direction. As for the
communication test, its p-values hovered around the boundary of significance, with the
two-sided test accepting the null hypothesis while the one-sided test rejected it. This
suggests a potential influence of ChatGPT on efficiency that warrants further investigation
with a larger sample size. Conversely, the informatics test exhibited a clear and significant
reduction in test time within the experimental group, indicating that ChatGPT significantly
enhanced test completion efficiency in this discipline.</p>
      <p>The results of the t-tests are attributed to the extent of ChatGPT usage and browser
usage. It should be highlighted that 26 students from the experimental group and 23 from
the control group could not be observed, in addition 70 students from the experimental
group chose to use ChatGPT. Within this, a total of 1518 prompts were observed, with the
middle 50% of students using between 10 and 32 prompts for the test. The prompts were
often a specific copying of the questions, with rewriting the prompt for an incorrect answer
being more common than rewriting the answer. On average, students spent 32.41 seconds
writing the prompt, reading the answer and regenerate it if needed.</p>
      <p>The distribution of ChatGPT usage across different test types reveals distinct patterns.
For the communication test, 72% of students did not use ChatGPT, while 28% did. In the
informatics test, 35% of students did not use ChatGPT, compared to 65% who did. For the
law test, the majority of students, 81%, did not use ChatGPT, with only 19% utilizing it. This
data highlights that ChatGPT usage was highest among informatics students and lowest
among law students. Students in business informatics used ChatGPT extensively, which
likely contributed to significantly better test times in this discipline. Focus time, indicating
how long a student was actively engaged with the browser, was also measured. For business
informatics, the average ChatGPT focus time was 6.45 minutes per test session, compared
to 2.39 minutes for law and 2.30 minutes for communication. This suggests that the
efficiency gains observed in business informatics may be attributed to the extensive use of
ChatGPT and the structured nature of the tasks in this field.</p>
      <p>The students who took the law and media and communication test in the experimental
group also preferred to browse, using Google. In the case of the experimental group of the
media and communication test, there was also little browsing activity, presumably
explaining the lower scores in the experimental group.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <sec id="sec-4-1">
        <title>4.1. Efficacy and effectiveness</title>
        <p>The discussion is concerning both the literature review and the results. Our analysis
indicates that while ChatGPT's integration into higher education research does not
significantly enhance the accuracy of student responses, it notably improves the efficiency
of completing tasks. This finding, endorsed by other researchers, suggests ChatGPT's
potential to boost student productivity by facilitating quicker research and problem-solving
[16], [28]. The varied impact of ChatGPT across disciplines underscores the importance of
a tailored approach to AI integration in education, cautioning against its unchecked use [19].
The varied impact of ChatGPT across disciplines highlights the need for a context-specific
approach to its integration. While it can streamline certain academic tasks, it's essential to
recognize its limitations in fostering critical thinking and creativity, which are vital for
comprehensive learning. This insight calls for a balanced integration strategy, positioning
AI tools as adjuncts to, rather than replacements for, traditional learning methods [17], [10].
Ensuring that AI tools enhance rather than diminish the development of key academic
competencies is crucial for their effective use in educational settings.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Personalized learning</title>
        <p>
          One of the standout benefits of integrating AI tools such as ChatGPT in educational settings
is the facilitation of personalized learning [26]. AI's adaptive learning technologies can
tailor educational content to meet the individual needs of students, considering their
learning pace, preferred learning styles, and areas of difficulty. This personalized approach
not only enhances student engagement but also addresses unique challenges, improving
overall learning outcomes. Furthermore, the use of ChatGPT and similar technologies in
education supports a more interactive and responsive learning environment. Unlike
traditional educational resources, these AI tools can provide instant feedback and
clarification, fostering a more dynamic and engaging learning experience [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The ability to
provide real-time adjustments to the learner's needs significantly contributes to the
effectiveness of the learning process, making learning more accessible and efficient.
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Overreliance on technology</title>
        <p>The dual potential of AI in education presents both opportunities for enhancement and risks
of dependency. In their paper, researchers highlight that reliance on AI for learning can lead
to a reduced development of critical thinking skills, even when AI explains its reasoning
[32]. This underscores the need for a balanced approach in utilizing AI tools to maintain the
integrity of educational processes. Educators must foster an environment where AI is used
to complement, not replace, critical engagement and independent problem-solving skills.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Lack of deep-learning and critical analysis.</title>
        <p>Reflecting on the challenges of implementing 'deep learning' in education, 'deep learning'
here means thorough understanding and mastery of information, not just superficial details
or problem-solving skills. It involves grasping core concepts, integrating new knowledge,
and applying it broadly. Literature highlights skepticism about AI’s role in fostering such
deep learning, raising concerns over reliance on AI-generated answers and superficial
content engagement [29]. Therefore, while AI tools like ChatGPT can offer quick
information, they should be used to support and enhance deep learning by helping students
gain a lasting understanding and effectively apply knowledge [22]. Educators should
leverage AI to add context and reinforce concepts, ensuring it aids in active, meaningful
learning.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Limitations</title>
      <p>
        Our study has several constraints. Firstly, the participant selection from various Hungarian
universities may limit the generalizability of our findings beyond this specific context [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
Secondly, variations in students' prior experiences with ChatGPT or Google, study habits,
and resource access could introduce biases that affect our result [18]. Additionally, the use
of the ExamEye browser extension may have caused the Hawthorne effect, potentially
altering students' genuine interactions with AI tools due to their awareness of being
observed [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Lastly, the short study duration may not capture ChatGPT's long-term effects
on educational outcomes, highlighting the need for future longitudinal research [30].
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion References</title>
      <p>This study assessed ChatGPT's impact on higher education students in law, business
informatics, and media and communication. While test scores did not improve, ChatGPT did
speed up test completion times, with varying effects across disciplines.
[9] Chiu, T. K. F. (2024). Future research recommendations for transforming higher
education with generative AI. Computers and Education: Artificial Intelligence, 6,
100197. https://doi.org/10.1016/j.caeai.2023.100197
[10] Cotton, D., Cotton, P A., &amp; Shipway, J R. (2023, March 13). Chatting and cheating:
Ensuring academic integrity in the era of ChatGPT. Innovations in Education and
Teaching International, 1-12. https://doi.org/10.1080/14703297.2023.2190148
[11] Curry, N., Baker, P., &amp; Brookes, G. (2024). Generative AI for corpus approaches to
discourse studies: A critical evaluation of ChatGPT. Applied Corpus Linguistics, 4(1),
100082. https://doi.org/10.1016/j.acorp.2023.100082
[12] Einarsson, H., Lund, S. H., &amp; Jónsdóttir, A. H. (2024). Application of ChatGPT for
automated problem reframing across academic domains. Computers and Education:
Artificial Intelligence, 6, 100194. https://doi.org/10.1016/j.caeai.2023.100194
[13] Else H. (2023). Abstracts written by ChatGPT fool scientists. Nature, 613(7944), 423.</p>
      <p>https://doi.org/10.1038/d41586-023-00056-7
[14] Essel, H. B., Vlachopoulos, D., Essuman, A. B., &amp; Amankwa, J. O. (2024). ChatGPT effects
on cognitive skills of undergraduate students: Receiving instant responses from
AIbased conversational large language models (LLMs). Computers and Education:
Artificial Intelligence, 6, 100198. https://doi.org/10.1016/j.caeai.2023.100198
[15] Firat M. (2023). What ChatGPT means for universities: Perceptions of scholars and
students. Journal of applied learning and teaching, 6(1).
https://doi.org/10.37074/jalt.2023.6.1.22
[16] García‐Martínez, I., Batanero, J M F., Fernández‐Cerero, J., &amp; León, S P. (2023, January
15). Analysing the Impact of Artificial Intelligence and Computational Sciences on
Student Performance: Systematic Review and Meta-analysis. Journal of New
Approaches in Educational Research, 12(1), 171-171.
https://doi.org/10.7821/naer.2023.1.1240
[17] Garrel, J V., &amp; Mayer, J. (2023, November 9). Artificial Intelligence in studies—use of
ChatGPT and AI-based tools among students in Germany. Humanities and Social
Sciences Communications, 10(1). https://doi.org/10.1057/s41599-023-02304-7
[18] Gopalan, M., Rosinger, K., &amp; Ahn, J. (2020). Use of Quasi-Experimental Research Designs
in Education Research: Growth, Promise, and Challenges. Review of Research in
Education, 44, 218-243.
[19] Grieve, R., Woodley, J., Hunt, S E., &amp; McKay, A. (2021, July 15). Student fears of oral
presentations and public speaking in higher education: a qualitative survey. Journal of
Further and Higher Education, 45(9), 1281-1293.
https://doi.org/10.1080/0309877x.2021.1948509
[20] Han, C., Feng, J., &amp; Qi, H. (2024). Topic model for long document extractive
summarization with sentence-level features and dynamic memory unit. Expert
Systems with Applications, 238(Part B), 121873.
https://doi.org/10.1016/j.eswa.2023.121873
[21] Liu, M., Zhang, L. J., &amp; Biebricher, C. (2024). Investigating students' cognitive processes
in generative AI-assisted digital multimodal composing and traditional writing.
Computers &amp; Education, 211, 104977.
https://doi.org/10.1016/j.compedu.2023.104977
[22] Liu, Y., Chen, L., &amp; Yao, Z. (2022). The application of artificial intelligence assistant to
deep learning in teachers' teaching and students' learning processes. Frontiers in
Psychology, 13.
[23] Luo, J. (Jess) (2024) A critical review of GenAI policies in higher education assessment:
a call to reconsider the “originality” of students’ work , Assessment &amp; Evaluation in
Higher Education, DOI: 10.1080/02602938.2024.2309963
[24] Miyazaki, K., Murayama, T., Uchiba, T. et al. Public perception of generative AI on
Twitter: an empirical study based on occupation and usage. EPJ Data Sci. 13, 2 (2024).
https://doi.org/10.1140/epjds/s13688-023-00445-y
[25] Mogavi, R H., Deng, C., Kim, J J., Zhou, P., Kwon, Y D., Metwally, A H S., Tlili, A., Bassanelli,
S., Bucchiarone, A., Gujar, S., Nacke, L E., &amp; Hui, P. (2024, January 1). ChatGPT in
education: A blessing or a curse? A qualitative study exploring early adopters’
utilization and perceptions. Computers in Human Behavior Artificial Humans, 2(1),
100027-100027. https://doi.org/10.1016/j.chbah.2023.100027
[26] Montebello, M. (2021). Personalized Learning Environments. 2021 International</p>
      <p>Symposium on Educational Technology (ISET), 134-138
[27] Moorhouse, B. L. (2024). Beginning and first-year language teachers' readiness for the
generative AI age. Computers and Education: Artificial Intelligence, 6, 100201.
https://doi.org/10.1016/j.caeai.2024.100201
[28] Nam, B H., &amp; Bai, Q. (2023, November 16). ChatGPT and its ethical implications for
STEM research and higher education: a media discourse analysis. International Journal
of STEM Education, 10(1). https://doi.org/10.1186/s40594-023-00452-5
[29] Perrotta, C., &amp; Selwyn, N. (2019). Deep learning goes to school: Toward a relational
understanding of AI in education. Learning, Media and Technology, 45(3), 251-269.
[30] Schneider, B., Saw, G., &amp; Broda, M. (2016). A Future for the National Education</p>
      <p>Longitudinal Program. AERA Open, 2.
[31] Stahl, B. C., &amp; Eke, D. (2024). The ethics of ChatGPT – Exploring the ethical issues of an
emerging technology. International Journal of Information Management, 74, 102700.
https://doi.org/10.1016/j.ijinfomgt.2023.102700
[32] Vasconcelos, H., Jörke, M., Grunde-McLaughlin, M., Gerstenberg, T., Bernstein, M., &amp;
Krishna, R. (2022). Explanations Can Reduce Overreliance on AI Systems During
Decision-Making. Proceedings of the ACM on Human-Computer Interaction, 7, 1-38.
[33] Wang, T., Diaz, D., Brown, C., &amp; Chen, Y. (2023, October 3). Exploring the Role of AI
Assistants in Computer Science Education: Methods, Implications, and Instructor
Perspectives. https://doi.org/10.1109/vl-hcc57772.2023.00018
[34] Zhai, X., Chu, X., Chai, C S., Jong, M S., Starčič, A I., Spector, M., Liu, J., Jing, Y., &amp; Li, Y. (2021,
April 20). A Review of Artificial Intelligence (AI) in Education from 2010 to 2020.</p>
      <p>Complexity, 2021, 1-18. https://doi.org/10.1155/2021/8812542
[35] Zhai, X., Nyaaba, M. &amp; Ma, W. Can Generative AI and ChatGPT Outperform Humans on
Cognitive-Demanding Problem-Solving Tasks in Science?. Sci &amp; Educ (2024).
https://doi.org/10.1007/s11191-024-00496-1</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Altmäe</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sola-Leyva</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Salumets</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Artificial intelligence in scientific writing: a friend or a foe?</article-title>
          <source>Reproductive BioMedicine Online</source>
          ,
          <volume>47</volume>
          (
          <issue>1</issue>
          ),
          <fpage>3</fpage>
          -
          <lpage>9</lpage>
          . https://doi.org/10.1016/j.rbmo.
          <year>2023</year>
          .
          <volume>04</volume>
          .009
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Baidoo-Anu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Owusu Ansah</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Education in the Era of Generative Artificial Intelligence (AI): Understanding the Potential Benefits of ChatGPT in Promoting Teaching and Learning</article-title>
          .
          <source>SSRN Electronic Journal.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bansal</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chamola</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hussain</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          et al.
          <article-title>Transforming Conversations with AI-A Comprehensive Study of ChatGPT</article-title>
          . Cogn
          <string-name>
            <surname>Comput</surname>
          </string-name>
          (
          <year>2024</year>
          ). https://doi.org/10.1007/s12559-023-10236-2
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Bhattacharya</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Nakhare</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2019</year>
          ,
          <article-title>November 1). Exploring AI-enabled Intelligent Tutoring System in the Vocational Studies Sector in UAE</article-title>
          . https://doi.org/10.1109/itt48889.
          <year>2019</year>
          .9075093
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Bower</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Torrington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>J.W.M.</given-names>
          </string-name>
          et al.
          <article-title>How should we change teaching and assessment in response to increasingly powerful generative Artificial Intelligence? Outcomes of the ChatGPT teacher survey</article-title>
          .
          <source>Educ Inf Technol</source>
          (
          <year>2024</year>
          ). https://doi.org/10.1007/s10639-023-12405-0
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Brown</surname>
          </string-name>
          , T. B.,
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Herbert-Voss</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krueger</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , ...
          <string-name>
            <surname>Amodei</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Language models are few-shot learners</article-title>
          . https://arxiv.org/abs/
          <year>2005</year>
          .14165
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Carminati</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Generalizability in Qualitative Research: A Tale of Two Traditions</article-title>
          .
          <source>Qualitative Health Research</source>
          ,
          <volume>28</volume>
          ,
          <fpage>2094</fpage>
          -
          <lpage>2101</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vander</surname>
            <given-names>Weg</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M. V.</given-names>
            ,
            <surname>Hofmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            , &amp;
            <surname>Reisinger</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          (
          <year>2015</year>
          ).
          <article-title>The Hawthorne Effect in Infection Prevention and Epidemiology</article-title>
          .
          <source>Infection Control &amp; Hospital Epidemiology</source>
          ,
          <volume>36</volume>
          (
          <issue>12</issue>
          ),
          <fpage>1444</fpage>
          -
          <lpage>1450</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>