<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Automatic Generation of Test Tasks Using ChatGPT API* *</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shcherbiak</string-name>
          <email>cherbiak@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tetiana Yurchyshyn</string-name>
          <email>tetianayurchyshyn@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tiande Pan</string-name>
          <email>pan54453@163.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Andriy Melnyk</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Catholic University in Ruzomberok Ruzomberok</institution>
          ,
          <addr-line>Hrabovská cesta 1A, Ruzomberok</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Ternopil Ivan Puluj National Technical University</institution>
          ,
          <addr-line>11 Lvivska Street, Ternopil, 46001</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>West Ukrainian National University</institution>
          ,
          <addr-line>11 Lvivska Street, Ternopil, 46001</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The development of effective and personalized assessment tools is a crucial challenge in contemporary education, particularly in the context of large-scale and digital learning environments. This paper presents a method for the automatic generation of test tasks using the ChatGPT API, a state-of-the-art language model developed by OpenAI. The proposed approach aims to reduce the workload of educators by automating the creation of diverse and pedagogically sound test items. The system enables the generation of questions tailored to various difficulty levels, ensuring alignment with students' individual knowledge profiles. Additionally, the method guarantees content uniqueness through dynamic question phrasing, thereby enhancing academic integrity and minimizing the reuse of static test materials. The flexibility of the ChatGPT API allows for the generation of multiple types of test questions, including multiple choice, short answer, and open-ended tasks across different subject areas. Experimental implementation demonstrates the feasibility of integrating this technology into modern educational platforms, resulting in improved efficiency, scalability, and personalization in the assessment process.</p>
      </abstract>
      <kwd-group>
        <kwd>Education system</kwd>
        <kwd>ChatGPT API</kwd>
        <kwd>test</kwd>
        <kwd>іnformation systems</kwd>
        <kwd>кnowledge testing methodology</kwd>
        <kwd>*</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the education system, the task of creating high-quality test questions is one of the key steps
in ensuring effective learning and assessment. However, this process is often time-consuming and
requires significant effort from teachers, methodologists, and developers. They must not only create
the tasks but also adapt them to the students’ knowledge levels, ensure the uniqueness of the
questions, and maintain content relevance [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1-4</xref>
        ].
      </p>
      <p>
        Automating the creation of test questions using artificial intelligence opens up new
opportunities for educational platforms and knowledge assessment systems. The use of language
models significantly reduces the time required to prepare tests, eliminates routine work, and
increases the productivity of the learning process. This is especially valuable in the context of mass
online education, where there is a need to generate thousands of unique test versions quickly [
        <xref ref-type="bibr" rid="ref5 ref6">5,6</xref>
        ].
      </p>
      <p>One of the key advantages of automated test creation is the ability to adjust the difficulty of
questions according to the user’s knowledge level. This makes it possible to generate personalized
test tasks that meet individual students' needs. As a result, the learning process becomes more
effective, and student engagement increases.</p>
      <p>
        Moreover, automatic test generation ensures the uniqueness of questions, which is particularly
important for preventing cheating and the reuse of the same questions. For instance, Student A might
receive the question: "What is the capital of France?", while Student B gets a similar but rephrased
question: "Name the main city of France." This approach makes it harder for students to share
readymade answers [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Modern artificial intelligence technologies, such as the ChatGPT API, enable the efficient
implementation of these capabilities. With this tool, it is possible to automatically generate different
types of test questions, control their difficulty, and tailor them to specific requirements. Thanks to
flexible settings, tests can be created for various disciplines and adapted accordingly [
        <xref ref-type="bibr" rid="ref7">7-10</xref>
        ].
      </p>
      <p>Thus, integrating the ChatGPT API into the test creation process significantly enhances the
quality and speed of preparation, ensuring both scalability and personalization. This makes artificial
intelligence a valuable assistant in the field of education and testing [8].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Main Capabilities and Limitations of the ChatGPT API</title>
      <p>
        Let’s take a closer look at what this tool is and how it can be used to automate the creation of
test questions. ChatGPT API is a programming interface that allows interaction with OpenAI’s
language model via HTTP requests. Through this interface, users can send text prompts and receive
generated responses, making it possible to automate numerous tasks, including the creation of test
questions, text analysis, generation of educational materials, and much more. The API is powered by
advanced natural language processing (NLP) algorithms, which ensures high quality and naturalness
of the generated texts [
        <xref ref-type="bibr" rid="ref7">7-9, 11</xref>
        ].
      </p>
      <p>One of the key features of the ChatGPT API is its ability to generate text based on a given
prompt. This makes it possible to create various types of questions: multiple-choice tests, open-ended
questions, situational tasks, crosswords, and even code snippets for assessing technical knowledge.
For example, you can send a request like: “Generate a physics test question for 10th grade with three
answer options,” and the API will return a corresponding text.</p>
      <p>Another important feature is support for contextual dialogue. This means the API can take
previous user messages into account during interaction, which is especially useful for adaptive
testing. For instance, if a student answers a question incorrectly, the system can generate additional
explanations or simplified questions to help them better understand the topic.</p>
      <p>Users can also adjust generation parameters to get the desired output. For example, the
temperature parameter controls the level of creativity in the responses: a low value (e.g., 0.2) makes
answers more predictable, while a higher value (0.8 –1.0) leads to more varied responses. The
max_tokens parameter limits the length of the response, helping to control the amount of text
generated.</p>
      <p>Although the ChatGPT API is a powerful tool, it does have certain limitations. First, there are
usage limits depending on the OpenAI pricing plan. For example, free or basic plans may have a
limited number of requests per month, which can be a significant factor for large-scale testing or the
automatic generation of large amounts of content.</p>
      <p>Second, the generated answers may be incorrect or require additional verification. Since
ChatGPT is based on statistical models, it does not guarantee 100% accuracy of factual information.
For example, when asked “What year was Isaac Newton born?”, the model might give the correct
answer (1643), but there is also a chance of inaccuracies. This means that generated test questions
should be reviewed, especially if they are used in formal education.</p>
      <p>Another important limitation is the lack of guaranteed uniqueness of test questions without
further processing. This means that repeated API calls with the same prompt may result in similar
or identical outputs. To avoid this, post-processing techniques can be used, such as randomly
rephrasing questions or combining elements from different generated variants.</p>
      <p>Despite these limitations, the ChatGPT API remains one of the most effective tools for automating
testing. For example, educational platforms can integrate the API into their systems to automatically
generate tests based on specific topics and difficulty levels. The API can also be used in corporate
training to assess employee knowledge, create quizzes, and test professional competencies.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Features of using the ChatGPT API</title>
      <p>
        To use the ChatGPT API, you need to have an OpenAI account. Registration takes place on the
official OpenAI website, after which the user gains access to a personal dashboard where they can
manage API keys. An API key is a unique code that allows applications to interact with the language
model via HTTP requests. Without this key, using the API is not possible. To work with the API, you
need a development environment and tools for executing HTTP requests. The most commonly used
setup includes Python with the requests or OpenAI library, as well as Postman for manual request
testing [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>Once everything is set up, you can perform API requests directly from your code. You can also
use Postman or cURL in the terminal to test API interaction before integrating it into a real project.
Figure 1 shows an example of a simple API request in Python.</p>
      <p>To control the behavior of the API, you can use additional parameters such as: temperature,
max_tokens, top_p, etc. Figure 2 shows an example of an API request in Python.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology for Generating Test Questions</title>
      <p>Generating test questions using the ChatGPT API requires a thorough approach, which includes
selecting the test format, properly formulating prompts, ensuring the quality of generated responses,
and automating the process. Thanks to the capabilities of the language model, it's possible to create
various types of tests, adjust their difficulty levels, and ensure the uniqueness of the questions. In
this section, we will take a detailed look at all the aspects necessary for the effective automatic
generation of test questions.</p>
      <p>Before getting started, it's important to determine which test format best suits your needs. The
most common types include:</p>
      <p>- multiple-choice questions, where the user selects the correct answer from a list of
options;
- open-ended questions, which require the user to formulate their own answer;
- logical problems, which assess analytical thinking skills;
- matching tests, where elements from two groups need to be paired;
- true/false tests, which allow for quick knowledge assessment;
- the choice of test type affects how prompts are constructed and how responses are
processed.</p>
      <p>Each test question should include a clearly formulated question, answer options (if it is a
multiple-choice question), the correct answer, and, if necessary, an explanation. For example, if we
are creating a test in JSON format for use in a testing system, it is advisable to include the following
structure (Fig. 3):</p>
      <p>This format ensures a standardized approach to organizing test questions, making it easy to
integrate them into testing systems. To get high-quality results from the API, it's important to
formulate prompts correctly. General requests like "Create a test question on Python" may lead to
vague or unstructured responses. Instead, you should use precise phrasing, such as: "Create a test
question on the topic 'Algorithms'. Format: question, 4 answer options, correct answer, short
explanation." Additionally, you can include parameters to control the difficulty level of the test, for
example: "Create a difficult test question on the topic 'Network Protocols' that requires in-depth
analysis."</p>
      <p>This approach helps generate well-structured and meaningful responses that can be easily
adapted to specific needs. To ensure that test questions match the knowledge level of the test-takers,
it's important to manage question difficulty. For example, three difficulty levels can be defined:
1) Beginner level – basic knowledge, e.g.: "What is a variable in programming?".
2) Intermediate level – questions that require analysis, e.g.: "What is the difference
between a list and a tuple in Python?"
3) Advanced level – questions that require an extended response, e.g.: "Describe how
the garbage collector works in Java."</p>
      <p>To receive a test question of the appropriate difficulty, you need to clearly specify it in the
API prompt: "Generate a test question for second-year students on the topic 'Databases'." One of the
useful features of the ChatGPT API is context retention, which allows the creation of cohesive test
sets. For example, you can first generate a list of key topics within a discipline and then create a
separate test question for each one. This approach improves the consistency of the test content and
helps avoid randomly generated questions that may not be relevant to the chosen topic.
Here’s an example sequence of prompts:</p>
      <sec id="sec-4-1">
        <title>1. "List 5 main topics from the course 'Network Technologies'."</title>
        <p>2. "Create one test question for each of these topics."</p>
        <p>This method results in more organized tests that align with the curriculum. If you need to
create a large set of test questions, the process can be automated. Using a simple Python script, you
can generate 10 test questions and save them to a JSON file (Fig. 4):</p>
        <p>This approach allows for the automatic generation of a large number of unique test questions,
which can be used for learning, certification, or knowledge assessment. Thanks to a well-structured
methodology, using the ChatGPT API for automatic test creation becomes an effective and reliable
tool. By following these principles, it's possible to ensure high-quality testing tailored to different
knowledge levels and educational programs.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Processing and Validation of Responses</title>
      <p>When automatically generating test questions using the ChatGPT API, it's important not
only to obtain the questions but also to ensure their quality, uniqueness, and alignment with the
expected difficulty level. Artificial intelligence can make mistakes or generate incorrect questions,
so an essential step is the validation of the generated test items. In this section, we’ll explore key
validation methods: verifying answer accuracy, checking for uniqueness, ensuring alignment with
educational standards, and using automated approaches to quality control.</p>
      <p>The first step after receiving a test question is to check whether the correct answer is indeed
correct. AI may occasionally generate an incorrect option or a poorly worded answer. Several
methods can be used for validation:
1. Manual review – An expert or instructor reviews the question and confirms the
correctness of the answers.
2. Cross-check via API – You can re-query ChatGPT with a prompt like: “Is this the correct
answer to the question?” to get confirmation.
3. Comparison with reliable sources – Validate the correct answers by comparing them
with documentation, textbooks, or academic resources.</p>
      <sec id="sec-5-1">
        <title>This process can be automated using an additional API request (Fig. 5):</title>
        <p>This approach helps filter out duplicate questions and ensures diversity in the tests. Even if
the test items are correct, it is important to assess whether they align with a specific educational
standard or curriculum. For example, beginner-level questions should not include complex
terminology, while advanced-level questions should cover complex concepts. To address this, a test
categorization system can be created:
● Basic level – definitions, simple terms
● Intermediate level – data analysis, syntax
● Advanced level – algorithms, system architecture</p>
        <p>This validation can also be performed using an automated approach by querying the API again
(Fig. 7):</p>
        <p>To significantly reduce the time required to validate a large number of test questions,
validation can be automated using a verification pipeline. For example, after a test is generated, it
can be automatically checked against the following criteria:</p>
      </sec>
      <sec id="sec-5-2">
        <title>1. Answer accuracy – Does the correct answer match a verified source?</title>
        <p>2. Uniqueness – Is the question already present in the database?
3. Difficulty level – Does the question match the intended educational level?
4. Objectivity – Does the question contain any bias or controversial information?
This can be implemented as a script (Fig. 9):</p>
        <p>This approach allows for automatically going through all stages of validation and obtaining
a structured assessment of the quality of test questions.</p>
        <p>Although automated validation can significantly simplify the process, final review should
still be performed by a human. Automated algorithms may not always recognize context or subtle
inaccuracies. The optimal approach is to combine automated filtering with selective review by
subject matter experts. This way, a balance between test generation speed and quality can be
achieved.</p>
        <p>Using these methods, it is possible to create a system that generates high-quality, unique, and
accurate test questions ready for use in educational or certification programs [8, 12-15].</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>Automating the creation of test questions using the ChatGPT API greatly simplifies the process
of developing educational assessments. The use of artificial intelligence allows for rapid generation
of diverse questions on any topic, customization of difficulty levels and answer formats, and
scalability for large groups of users. However, despite the model’s powerful capabilities, it is
important to ensure the quality of generated questions by verifying their accuracy and uniqueness,
and by filtering out incorrect or biased phrasing. The optimal solution is a combination of automated
generation and manual review, which enables the creation of reliable and effective tests.</p>
      <p>The potential applications of this approach are vast. Test generation can be integrated into
learning management systems, certification platforms, and automated knowledge assessment tools.
With the ability to dynamically modify test content, it is possible to adapt tests to individual users,
creating personalized tasks that reflect their knowledge level and progress. The ChatGPT API can
also be used alongside test result analytics to enhance educational programs.</p>
      <p>Declaration on Generative AI
During the preparation of this work, the authors used ChatGPT and Grammarly to check grammar
and spelling, paraphrase, and reword the text. These tools help identify and correct grammatical
errors, typos, and other writing mistakes, improving the clarity and professionalism of the text. After
using these tools, the authors reviewed and edited the content as needed and take full responsibility
for the publication’s content.
[8] Silva, Carlos Alexandre Gouvea da, Felipe Negrelle Ramos, Rafael Veiga de Moraes, and Edson
Leonardo dos Santos. 2024. "ChatGPT: Challenges and Benefits in Software Programming for
Higher Education" Sustainability 16, no. 3: 1245. https://doi.org/10.3390/su16031245
[9] Coello, Carlos Eduardo Andino, Mohammed Nazeh Alimam, and Rand Kouatly. 2024.
"Effectiveness of ChatGPT in Coding: A Comparative Analysis of Popular Large Language
Models" Digital 4, no. 1: 114-125. https://doi.org/10.3390/digital4010005
[10] Karnalim, O., &amp; Kurniawati, G. (2020). PROGRAMMING STYLE ON SOURCE CODE
PLAGIARISM AND COLLUSION DETECTION. International Journal of Computing, 19(1),
2738. https://doi.org/10.47839/ijc.19.1.1690 PLAGIARISM AND COLLUSION DETECTION.</p>
      <p>International Journal of Computing, 19(1), 27-38. https://doi.org/10.47839/ijc.19.1.1690
[11] Palagin, O., Kaverinskiy, V., Litvin, A., &amp; Malakhov, K. (2023). OntoChatGPT Information
System: Ontology-Driven Structured Prompts for ChatGPT Meta-Learning. International
Journal of Computing, 22(2), 170-183. https://doi.org/10.47839/ijc.22.2.3086
[12] Khomytska, I., Teslyuk, V., Bazylevych, I., &amp; Shylinska, I. (2020). APPROACH FOR
MINIMIZATION OF PHONEME GROUPS IN AUTHORSHIP ATTRIBUTION. International
Journal of Computing, 19(1), 55-62. https://doi.org/10.47839/ijc.19.1.1693
[13] A. Melnyk, I. Shcherbiak, R. Shevchuk, A. Shevchuk, O. Huhul and Y. Franko, "Software
Architecture for Mathematical Modelling Based on Interval and Ontology Approach," 2022 12th
International Conference on Advanced Computer Information Technologies (ACIT),
Ruzomberok, Slovakia, 2022, pp. 101-105, doi: 10.1109/ACIT54803.2022.9913108.
[14] N. Pasichnyk, A. Melnyk and N. Dobrovolska, "Management the website attendance based on
the projected traffic," 2013 12th International Conference on the Experience of Designing and
Application of CAD Systems in Microelectronics (CADSM), Lviv, UKraine, 2013, pp. 277-277.
[15] Gordillo, Aldo. 2019. "Effect of an Instructor-Centered Tool for Automatic Assessment of
Programming Assignments on Students’ Perceptions and Performance" Sustainability 11, no. 20:
5568. https://doi.org/10.3390/su11205568</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Novak</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kermek</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Assessment Automation of Complex Student Programming Assignments</article-title>
          .
          <source>Educ. Sci</source>
          .
          <year>2024</year>
          ,
          <volume>14</volume>
          , 54. https://doi.org/10.3390/educsci14010054
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pasichnyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Melnyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pasichnyk</surname>
          </string-name>
          and
          <string-name>
            <surname>I. Turchenko</surname>
          </string-name>
          ,
          <article-title>"Method of adaptive control structure learning based on model of test's complexity,"</article-title>
          <source>Proceedings of the 6th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems</source>
          , Prague, Czech Republic,
          <year>2011</year>
          , pp.
          <fpage>692</fpage>
          -
          <lpage>695</lpage>
          , doi: 10.1109/IDAACS.
          <year>2011</year>
          .
          <volume>6072858</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Maslovskyi</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Sachenko</surname>
          </string-name>
          ,
          <article-title>"Adaptive test system of student knowledge based on neural networks,"</article-title>
          <source>2015 IEEE 8th International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS)</source>
          , Warsaw, Poland,
          <year>2015</year>
          , pp.
          <fpage>940</fpage>
          -
          <lpage>944</lpage>
          , doi: 10.1109/IDAACS.
          <year>2015</year>
          .7341442
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Maslovskyi</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Sachenko</surname>
          </string-name>
          ,
          <article-title>"Adaptive test system of student knowledge based on neural networks,"</article-title>
          <source>2015 IEEE 8th International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS)</source>
          , Warsaw, Poland,
          <year>2015</year>
          , pp.
          <fpage>940</fpage>
          -
          <lpage>944</lpage>
          , doi: 10.1109/IDAACS.
          <year>2015</year>
          .7341442
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Hadzhikoleva</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Rachovski,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Ivanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            ;
            <surname>Hadzhikolev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ;
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Automated Test Creation Using Large Language Models: A Practical</surname>
          </string-name>
          <article-title>Application</article-title>
          .
          <source>Appl. Sci</source>
          .
          <year>2024</year>
          ,
          <volume>14</volume>
          , 9125. https://doi.org/10.3390/app14199125
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Lendyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rippa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Bodnar</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Sachenko</surname>
          </string-name>
          ,
          <article-title>"Ontology Application in Context of Mastering the Knowledge for Students,"</article-title>
          <source>2018 IEEE 13th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT)</source>
          , Lviv, Ukraine,
          <year>2018</year>
          , pp.
          <fpage>123</fpage>
          -
          <lpage>126</lpage>
          , doi: 10.1109/STC-CSIT.
          <year>2018</year>
          .
          <volume>8526710</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Roumeliotis</surname>
            ,
            <given-names>Konstantinos I.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Nikolaos</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Tselikas</surname>
          </string-name>
          .
          <year>2023</year>
          .
          <article-title>"ChatGPT and Open-AI Models: A Preliminary Review" Future Internet 15</article-title>
          , no.
          <volume>6</volume>
          : 192. https://doi.org/10.3390/fi15060192
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>