<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Knowledge engineering information technology for cultural-educational scenarios based on RAG</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Khrystyna Lipianina-Honcharenko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nazar Melnyk</string-name>
          <email>88nazar88@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Myroslav Komar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavlo Bykovyy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Khrystyna Yurkiv</string-name>
          <email>kh.yurkiv@wunu.edu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>West Ukrainian National University</institution>
          ,
          <addr-line>Lvivska 11, 46000 Ternopil</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The digital transformation of cultural heritage and education necessitates systems capable of reliably extracting knowledge from large-scale corpora while presenting it in an understandable, verifiable, and ethically safe manner. However, the application of Large Language Models (LLMs) in museum scenarios is often hindered by hallucinations, a lack of source traceability, and safety risks for child audiences. In this work, we propose a reproducible knowledge engineering information technology based on Retrieval-Augmented Generation (RAG), specifically designed for interactive “talking exhibits.” The proposed architecture integrates ethical corpus construction, semantic search with re-ranking, and controlled generation with explicit source citation and “honest uncertainty” policies. Validation was conducted on a demonstration case of the historical figure Jan Tarnowski, where the system achieved 100% compliance with role-based linguistic constraints (8/8 queries) and 62.5% contextual faithfulness (5/8 queries fully confirmed). While localized risks of hallucinations persisted in specific domains such as law and finance, the outcomes demonstrate a balanced improvement in accuracy, explainability, and safety compared to isolated LLM usage. These results suggest that the proposed technology efectively bridges the gap between open generative models and the rigorous requirements of educational cultural heritage applications.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Retrieval-augmented generation</kwd>
        <kwd>cultural heritage</kwd>
        <kwd>knowledge engineering</kwd>
        <kwd>semantic search</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The digital transformation of cultural heritage and education has highlighted the need for systems
capable of reliably extracting knowledge from large-scale open corpora and presenting it in an
understandable, verifiable, and ethically safe form for diverse audiences. Despite the advances of large
language models (LLMs), their application in museum and cultural-educational scenarios faces several
limitations: from hallucinations and lack of source traceability to licensing risks and multilingual
challenges [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. The absence of reproducible end-to-end methodologies that simultaneously
combine ethical corpus construction, semantic search, controlled generation with citation, and evaluation
protocols underlines the scientific and practical significance of this research.
      </p>
      <p>The aim of this work is to develop and verify a RAG-based (Retrieval-Augmented Generation)
knowledge engineering technology for cultural-educational applications that:
1. builds a licensing-compliant and reproducible knowledge corpus;
2. ensures semantic search and re-ranking of relevant fragments;
3. generates grounded responses with explicit source citation;
4. implements safety policies and personal data minimization;
5. supports telemetry and ofline/online quality evaluation.</p>
      <p>
        The practical focus is on interactive museum “talking exhibits” that respond in the first person,
mirroring the user’s language of interaction, and demonstrating fact traceability. The studied problem is
interdisciplinary: it requires coordinated solutions at the levels of data collection and cleaning (HTML
cleaning, normalization, deduplication, versioning), information retrieval (vectorization, indexing,
reranking), natural language processing (controlled generation, role-based modeling, style/tone control),
and safety engineering (content moderation, child-audience policies, privacy) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. A particular challenge
lies in ensuring traceability: each generated fact must be linked to specific corpus fragments, which
increases user trust and facilitates expert verification.
      </p>
      <p>The working hypothesis is that combining a licensing-compliant corpus, semantic search with
re-ranking, and controlled generation with citation achieves a better balance between accuracy,
explainability, and safety than isolated use of LLMs or traditional search methods. We expect that the
proposed technology will provide stable language mirroring, first-person narrative, increased contextual
faithfulness, and controlled behavior in cases of uncertainty (refusal/clarification instead of fabrication).</p>
      <p>The paper is structured as follows. Section 2 systematizes approaches to ethical corpus construction
and RAG systems. Section 3 presents the architecture and step-by-step protocol. Section 4 outlines the
pilot and validation results, followed by conclusions in Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>The literature review in this work structures contemporary approaches to building RAG systems for
cultural-educational scenarios through five interrelated vectors:
1. ethical corpus construction (HTML cleaning, normalization, deduplication, licensing compliance,
and attribution);
2. methods of text segmentation into semantic fragments (fixed-size, structure-oriented, semantic,
and hierarchical chunking) and their vectorization for semantic search;
3. re-ranking and context compression considering positional and semantic features;
4. controlled generation with explicit source citation, style/tone control, and “honest uncertainty”
policies;
5. safety and audience compliance (including child audiences), as well as ofline/online evaluation.</p>
      <p>
        This perspective makes it possible to align the requirements of traceability and reproducibility with
the practical applicability of museum “talking exhibits.” At the same time, several gaps are identified [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]:
the lack of standardized fragment-level attribution protocols, insuficient coverage of multilingual
corpora, and limited development of metrics sensitive to factual groundedness.
      </p>
      <p>
        For building such systems, open text data are typically used—web pages, archives, and other sources.
A notable example is the C4 corpus (Colossal Clean Crawled Corpus), created on the basis of Common
Crawl with subsequent cleaning [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. During HTML parsing, libraries such as Boilerpipe or Readability
are employed to extract the main text and metadata while filtering out advertisements or menus.
However, filtering may have unintended consequences: for instance, in C4, the use of banned-word
lists led to the removal of texts about marginalized groups. Consequently, more cautious methods are
now applied—combining dictionaries, toxic content detection, and classifiers for identifying personal
data. In addition to cleaning, text normalization is essential: correcting encodings, merging split words,
and standardizing case and spelling variants.
      </p>
      <p>
        License and usage verification is also a critical step. In the research community, the movement
toward open data has accelerated. The Common Pile v0.1 corpus, comprising 8 TB of data, contains
only texts with open licenses (public domain, Creative Commons) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Results showed that language
models trained on such an “ethical” corpus achieve quality comparable to models trained on unverified
web data [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Therefore, in educational and cultural projects, it is recommended to use sources with
explicit licenses (e.g., Wikipedia, open-access publications) or to obtain permission from rights holders,
as well as to ensure proper attribution of sources in generated responses.
      </p>
      <p>
        Content deduplication is another key stage. During web scraping, duplicates or near-duplicate texts
often appear. In the C4 dataset, for example, a 50-word passage was found repeated 60,000 times [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
The study [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] demonstrated that removing duplicates accelerates model training (by reducing dataset
size) and lowers the risk of memorization—cases where the model reproduces training text verbatim.
Deduplication algorithms range from simple string hashing to building sufix arrays and searching for
repeated segments with a predefined similarity threshold [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        For efective information retrieval, large texts are split into semantic fragments (chunks). Diferent
chunking strategies are discussed in current research. The simplest approach is fixed-size segmentation
(e.g., 500 characters per fragment). While easy to implement, this may split logically connected
paragraphs [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Another approach is recursive segmentation by structural elements (newlines, punctuation),
which takes into account text formatting. A more advanced method is semantic chunking, which
groups sentences based on embedding similarity. Semantic chunks are usually internally coherent,
but overly narrow grouping may lose the broader context of the document. Recent studies propose
hybrid solutions, such as hierarchical segmentation, where smaller segments are first formed and then
clustered into higher-level units by meaning [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        The LongRAG model [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] uses larger fragments (entire articles) to reduce the size of the knowledge
base, while the RAPTOR method [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] builds multi-level hierarchies ranging from detailed content to
higher-level abstractions. All of these approaches aim to preserve coherent semantic units, thereby
making retrieval more accurate and useful.
      </p>
      <p>
        After segmentation, each chunk is converted into a vector using embedding models (e.g.,
SentenceBERT or large language models in embedding mode). Vectorization of knowledge enables semantic
search: the user query is also transformed into a vector, and the system retrieves the closest chunk
vectors using cosine similarity. For the implementation of vector knowledge bases, the FAISS (Facebook
AI Similarity Search) library is widely used [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. This toolkit provides fast Approximate Nearest
Neighbor (ANN) search for billions of embeddings and is integrated into vector databases such as Milvus
and Pinecone.
      </p>
      <p>
        Semantic search based on embeddings enables retrieval of relevant fragments, but the nearest vector
by distance does not always yield the best context. Therefore, many approaches employ additional
re-ranking of results. A classical method is to pass the top-K retrieved chunks through a stronger
model (e.g., a BERT cross-encoder or an LLM) for more accurate relevance scoring. It is well known
that BERT-based re-rankers significantly improve performance in text-based Question Answering (QA)
tasks compared to sparse retrieval or embedding-only retrieval [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
        Other optimizations include position-sensitive ranking, where keyword proximity to the beginning
of a chunk is considered, and context compression, where non-essential details are discarded from
candidate fragments [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. In domain-specific systems, context merging may also be used—when multiple
chunks refer to the same entity, they can be combined into a joint context before answer generation.
However, overly aggressive merging may lead to errors.
      </p>
      <p>
        The efectiveness of retrieval is evaluated using standard information retrieval metrics—Precision@K,
Recall@K, MRR, NDCG, etc.—to determine whether the top-ranked documents contain the “correct”
fragments [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. For greater accuracy, the Context Recall metric is sometimes applied, which measures
whether the retrieved context includes all facts from the reference answer. During system development,
engineers employ a set of control queries with known answers to verify whether the retrieval module
can locate the required paragraph.
      </p>
      <p>
        Once relevant fragments are retrieved, a large language model (LLM) generates the response. This
approach, which leverages external knowledge, is called retrieval-augmented generation (RAG). It
significantly improves the accuracy and factual reliability of responses while reducing model
hallucinations [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. In particular, the LLM receives a system prompt or instruction to incorporate the retrieved
facts and avoid unsupported claims. The best results are achieved by prompting with context, i.e.,
inserting knowledge fragments directly into the model’s input prompt, which ensures that the answer
is grounded in the provided sources. For example, the OpenAI WebGPT project demonstrated that
when a model cites web sources to support its answers, users perceive it as more trustworthy [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        An alternative approach is post-hoc citation, where the answer is generated first and then aligned
with corresponding sources. However, in practice, one-step methods are more common, where the
model integrates citations directly into the generated text. This is implemented, for instance, in the
Atlas system, which attaches a document index to each fact [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        Scientific research also addresses the control of style and tone in generated responses. For educational
applications—especially in child mode—the model should answer at a level comprehensible to children
and consistent with ethical norms. Two approaches are possible: (i) training a dedicated child-focused
model, or (ii) adapting a universal model with style-specific instructions. The KidLM project
demonstrated the advantages of the first approach: a dedicated children’s corpus ( ∼ 50 million words) was
collected, and the language model was further trained, leading to better simplification of complex
vocabulary and avoidance of toxic expressions [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Alternatively, when using a general-purpose LLM,
roles or modes can be embedded via prompting. Such techniques belong to prompt engineering and are
widely applied in practice.
      </p>
      <p>Content safety is a critical factor for cultural-educational applications. It is well known that large
language models may generate undesirable or harmful text if prompted, or if such content was present
in training data. The authors of “LLM Safety for Children” developed specialized child personas and
scenarios for testing several state-of-the-art models and identified categories of risks that are not
reliably avoided by standard filters [ 17]. This underscores the necessity of specialized filters and
rule-based safeguards. In existing solutions, moderation is typically implemented at two levels. The
ifrst is query-level filtering, where the system either refuses or provides a safe response to prohibited
topics. The second is answer-level filtering, where toxicity detectors evaluate the generated response
before presenting it to the user. The tone of responses is also regulated: the model should be friendly,
encourage curiosity, yet avoid harmful advice. The study “No, Alexa, no!” [18] highlighted the risk of
children placing excessive trust in voice assistants and the need to impose restrictions so that AI does
not present itself as an authoritative adult in sensitive topics.</p>
      <p>
        To maintain high system quality, both ofline and online evaluation are conducted. System telemetry
(Stage 11) makes it possible to collect statistics on dialogue counts, popular queries, and changes in
accuracy after updates. These data are typically used in industrial deployments for iterative improvement
(e.g., A/B testing of prompts, comparing model versions). In scientific publications, however, the focus
is placed on objective metrics. Specifically, the quality of knowledge retrieval is assessed using metrics
such as Recall@5 or MRR, which indicate how well the system retrieves the correct documents [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>Alongside classical information-retrieval measures, modern RAG research increasingly relies on
metrics that explicitly assess whether generated claims are grounded in retrieved context. RAGAS
(Retrieval-Augmented Generation Assessment) and ARES (Automatic RAG Evaluation System) represent
state-of-the-art frameworks combining retrieval checks with LLM-based critics to score groundedness,
hallucination frequency, and answer faithfulness to evidence [19, 20]. Unlike lexical metrics such as
BLEU/ROUGE, these approaches focus on semantic alignment and factual traceability, providing a more
reliable evaluation of context-dependent systems.</p>
      <p>System behavior is also tested under conditions of knowledge scarcity. The correct approach is for
the system to acknowledge uncertainty or ofer alternatives (e.g., suggest additional search). In research,
this issue is termed refusal or abstention. The R-Tuning (Refusal-aware tuning) method trains LLMs to
abstain from fabrications when a query falls outside their knowledge boundaries [21]. Models trained
in this way better recognize uncertainty and answer with “I don’t know” instead of hallucinating. For
educational applications, this is especially important: when uncertain, the system should direct users to
an expert or state that more information is needed, rather than providing incorrect guidance.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Technology description</title>
      <p>A knowledge engineering technology has been developed for cultural-educational scenarios that
includes:
1. collection and normalization of text sources;
2. segmentation (chunking) with overlap;
3. vectorization (embedding) and indexing;
4. semantic search and context re-ranking;
5. controlled LLM-based answer generation with grounding guarantees;
6. telemetry, ofline/online evaluation, and testing protocol.</p>
      <p>The technology is suitable for both “adult” and “child” modes, adheres to principles of personal data
minimization, and ensures reproducibility. The proposed technology consists of 13 stages, as illustrated
in the diagram (Figure 1).</p>
      <p>The detailed stages are as follows:
1. Collection and Parsing of Open Sources:
a) Automatically or semi-manually collect links.
b) Download pages as raw bytes, handling redirects and errors.</p>
      <p>c) Extract clean text from HTML along with basic metadata (title, source, date).
2. Verification of Usage Rights and Attribution:
a) Record license terms, copyrights, and citation requirements.</p>
      <p>b) Exclude sources that do not permit reuse.
3. Cleaning, Normalization, and Content Filtering:
a) Fix encodings, remove extraneous blocks such as menus and advertisements.
b) Normalize spacing, line breaks, languages, and transliterations.</p>
      <p>c) Remove toxic, undesirable, or personally sensitive content.
4. Deduplication and Corpus Versioning [22, 23]:
a) Detect duplicates by textual similarity.</p>
      <p>b) Store control hashes, corpus version, and change logs.
5. Segmentation into Semantic Fragments [22, 23]:
a) Create chunks respecting paragraphs and headings.
b) Apply overlapping windows to improve context retrieval.</p>
      <p>c) Add service tags (section, source, time period).
6. Knowledge Vectorization:
a) Compute embeddings for each chunk.
b) Normalize and store arrays for eficient cosine similarity search.</p>
      <p>c) Validate baseline quality with control queries.
7. Preparation of Storages and Indexes:
a) Store texts and vectors in a repository (file or database).
b) Record the vector file path in the exhibit database.</p>
      <p>c) Configure backups and integrity checks.
8. Exhibit Configuration in the Admin Panel:
a) Create an exhibit record (title, description, image, system settings).
b) Indicate whether child mode is enabled.</p>
      <p>c) Link the path to the vector file.
9. Safety and Moderation Policies:
a) Define rules for style, tone, and restricted topics.
b) Add instructions for child audiences (simplified language, caution with sensitive topics).
c) Enable baseline filters and answer checks before display.
10. Client Flow Configuration:
a) Session start form with minimal data (date of birth, anonymous identifier).
b) Branching by age (child or standard mode).
c) Construct a message for the language model (system prompt, user question, retrieved
contexts).
11. Analytics and Logging:
a) Log session starts, number of messages, average dialogue frequency, errors, and timings.
b) Prepare summary statistics for the admin panel.
12. Knowledge Retrieval Testing:
a) Use a set of control queries that must retrieve specific fragments.</p>
      <p>b) Evaluate context relevance, coverage accuracy, and retrieval stability.
13. Model Answer Evaluation:
a) Verify factual correctness, completeness, style, and tone alignment.
b) Test behavior under knowledge gaps—the model should honestly indicate lack of
information.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Implementation</title>
      <p>To construct the demonstration knowledge corpus, automated collection and preprocessing of open
text sources about the historical figure Jan Tarnowski were performed. At the search-and-parsing
stage, 8 unique links were identified; during main content extraction, 7 resources were successfully
processed, while 1 resource returned an error (details recorded in the execution log). The combined text
was consolidated into a single file /content/tarnowski_corpus.txt, and process metadata were
saved in /content/tarnowski_log.jsonl.</p>
      <p>The pipeline included encoding normalization, HTML cleaning, extraction of article “body” text,
denoising, and basic deduplication. The resulting material was segmented into overlapping semantic
fragments for subsequent semantic search. After corpus formation, the text file was uploaded to
the developed platform (Figure 2), where an exhibit was created in the admin module and a vector
representation of knowledge was generated for approximate fragment retrieval. This integration ensures
reproducibility, quality control (via state and encoding logs), and readiness for initial expert testing of
functionality, independent of a specific museum object.</p>
      <p>Next, the avatar’s role-based linguistic behavior was configured so that the character responds in
the first person and in the language of the user’s query, relying exclusively on the uploaded corpus.
The system prompt (see Figure 3) defines the formal role model of the exhibit’s linguistic behavior and
regulates the boundaries of generation. Its design aims to minimize hallucinations, preserve historical
stylistics, and maintain controllable response tone for diferent audiences.</p>
      <p>The prompt fixes the narrator’s identity (first-person perspective), ensures language mirroring
(responding in the user’s language), enforces strict linkage to the source corpus (only the provided
museum context), and applies an uncertainty policy (honest refusal in case of missing data). A response
length limit of five sentences disciplines the output and reduces the risk of accumulating secondary
assumptions.</p>
      <p>Next, a simple validation test of the avatar was conducted using queries (five in Ukrainian and one
each in English, Polish, and German) aimed at verifying first-person narrative, language mirroring, and
grounding in the provided corpus. The list of queries is presented in Table 1.</p>
      <p>The results of the mini-test are recorded in Table 2 (Figure 4) according to four criteria — compliance
with the query language, adherence to first-person narrative, grounding in the corpus, and response</p>
      <p>Розкажи про своє походження та раннi роки: з якого дому First-person narrative, basic
bioти походиш i де народився? graphical facts from the corpus
(EN) Tell me about your origin and early years: from which
family do you come, and where were you born?
Що саме ти заснував у Тарнополi й з якою метою? Factual accuracy, local event
(EN) What exactly did you establish in Ternopil, and for what (foundation)
purpose?
Якi вiйськовi кампанiї або походи ти очолював i якi наслiд- Historical events without
fabricaки вони мали? tion, concise summary
(EN) Which military campaigns or expeditions did you lead,
and what were their outcomes?
Якi лицарськi чесноти ти вважав найважливiшими для Consistency with style and values
себе? in the corpus
(EN) Which knightly virtues did you consider most important
for yourself?
Коли i де завершилося твоє життя, i як пройшли поховальнi Chronology, verification of corpus
урочистостi? references
(EN) When and where did your life end, and how were the
funeral ceremonies conducted?
Якими першоджерелами у наданому корпусi згадуються Works with “sources”; language
твої укази чи листування, i що з них випливає? mirroring (UA)
(EN) Which primary sources in the provided corpus mention
your decrees or correspondence, and what do they reveal?
Jak twoje decyzje gospodarcze i zarządzanie majątkiem wpły- Economic/administrative
influnęły na rozwój regionu? ence; language mirroring (PL)
(EN) How did your economic decisions and estate
management influence the development of the region?
Wie wird dein Andenken in späteren Chroniken und Studien Image in later historiography;
lanbeschrieben? guage mirroring (DE)
(EN) How is your memory described in later chronicles and
studies?</p>
      <p>Original Answer: Я, Ян Тарновський, належу до давнього магнатського дому Тарновських (гербу
Лелiва) з Малої Польщi...</p>
      <p>Translation: I, Jan Tarnowski, belong to the ancient magnate house of Tarnowski...</p>
      <sec id="sec-4-1">
        <title>Language (+) / First-person (+) / Response time (s) 15.1/9.6</title>
        <p>Context faithfulness: High. Confirmed: “Leliwa,” “Castellan of Kraków,” “Tarnów,” “Maciej Drzewiecki,”
references to classical education (Virgil)
Original Answer: Я заснував у 1540 роцi на Волинi мiсто–фортецю Тарнопiль (назване на честь
мого родового герба i Тарнова)...</p>
        <p>Translation: I founded in 1540, in Volhynia, the fortress-town of Tarnopol...</p>
      </sec>
      <sec id="sec-4-2">
        <title>Language (+) / First-person (+) / Response time (s) 9.3/9.6</title>
        <p>Context faithfulness: High. Confirmed: “1540,” “Tarnopol/Ternopil,” “Magdeburg rights,” defensive
function (Tatars).</p>
        <p>Original Answer: Я очолював двi найвiдомiшi воєннi кампанiї. У 1531 роцi, за наказом короля
Сигiзмунда...</p>
        <p>Translation: I led two of the most renowned military campaigns. In 1531, by order of King Sigismund...</p>
      </sec>
      <sec id="sec-4-3">
        <title>Language (+) / First-person (+) / Response time (s) 13.4/9.6</title>
        <p>Context faithfulness: High. Confirmed: “Obertyn/1531,” “Khotyn/1538,” “Petru Rares, .”
Original Answer: Найважливiшими для себе я вважав передусiм справедливiсть i
помiркованiсть...</p>
        <p>Translation: The virtues I considered most important for myself were, above all, justice and
moderation...</p>
      </sec>
      <sec id="sec-4-4">
        <title>Language (+) / First-person (+) / Response time (s) 12.5/9.7</title>
        <p>Context faithfulness: Medium-High. Present in the corpus are the virtues: justice, moderation, mercy,
piety, humility; the formulation generalizes but does not contradict.</p>
        <p>Original Answer: Моє життя завершилося 16 травня у 73-му роцi вiку. Помер я в своїх володiннях...
Translation: My life ended on May 16th, in the seventy-third year of my age...</p>
      </sec>
      <sec id="sec-4-5">
        <title>Language (+) / First-person (+) / Response time (s) 20.0/9.8</title>
        <p>Context faithfulness: High. Present are “May 16,” “73,” “Tarnów,” “Bernardines,” “bishops”; the
description of the funeral is confirmed by corpus fragments.</p>
        <p>Original Answer: In the royal court of Piotrków you may still read the very edicts I had nailed upon
the doors...</p>
      </sec>
      <sec id="sec-4-6">
        <title>Language (+) / First-person (+) / Response time (s) 17.1/10.1</title>
        <p>Context faithfulness: Low (partial). The corpus does not contain «Пйотркув» and «Янош» in the
relevant context... specific details about “прибитi едикти” are not confirmed.</p>
        <p>Original Answer: Jako gospodarz dóbr nie gromadziłem skarbów dla samego siebie...
Translation: As the steward of my estates, I did not amass treasures for myself...</p>
      </sec>
      <sec id="sec-4-7">
        <title>Language (+) / First-person (+) / Response time (s) 16.1/10.3</title>
        <p>Context faithfulness: Low. No links found for “Piotrków + two grosze”; financial details are not
confirmed.</p>
        <p>Original Answer: Ich bitte um Entschuldigung, doch in den mir vorliegenden Quellen findet sich keine
Angabe darüber...</p>
        <p>Translation: I apologize, but in the sources available to me, there is no information...</p>
      </sec>
      <sec id="sec-4-8">
        <title>Language (+) / First-person (+) / Response time (s) 7.9/10.4</title>
        <p>Context faithfulness: Neutral/cautious. Correct refusal without assertions; no explicit evidence
regarding “later historiography.”</p>
        <p>Linguistic-role behavior (see Table 2) was stable: compliance with the query language — 8/8 (100%),
adherence to the first-person narrative — 8/8 (100%). The answers maintained the prescribed style and
tone, contained no deviations from the “I-narrative,” and correctly mirrored the query language, which
confirms the efectiveness of the system prompt and mode routing.</p>
        <p>The context fidelity rate was 5/8 (62.5%) responses with full or predominant factual confirmation from
the corpus (Q1–Q5), 2/8 (25%) contained unconfirmed details (Q6–Q7), and 1/8 (12.5%) represented a
correct cautious refusal (Q8). Confirmed markers included “Leliwa,” “Tarnów,” “Obertyn 1531,” “Khotyn
1538,” “Petru Rares, ,” “16 May,” “Bernardines,” “bishops.” The lack of confirmation in Q6–Q7 was due to
the absence in the corpus of specific mentions such as “Piotrków,” “Janus/János,” “two groszy per lan for
veterans,” indicating a risk of localized hallucinations in the domains of law/finance/titulature.</p>
        <p>In terms of temporal characteristics, the actual mean response time was 13.93 s versus the forecast of
9.89 s (average delta +4.04 s; median 14.25 s vs. 9.75 s;  = 3.76 s). Recommendations for further piloting
include maintaining low stochasticity (temperature ≤ 0.2), increasing top_k to 5 with a relevance
threshold, limiting output length (≤ 4 sentences), if necessary separating diferent “personae” into
distinct exhibits, and supplementing the corpus with verified sources on problematic topics.</p>
        <p>Overall, the system demonstrates high compliance with the prescribed linguistic role and acceptable
accuracy on core corpus facts (62.5% supported), but requires localized adjustments of prompts/corpus
to mitigate hallucination risks in domain-specific responses.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>This study developed and verified a knowledge engineering information technology based on
RetrievalAugmented Generation (RAG) for cultural and educational applications. The stated objectives were
achieved: a reproducible and license-clean pipeline for corpus construction was built, semantic search
with re-ranking was implemented, and grounded answer generation with explicit citation and safety
mechanisms was ensured. Quantitative results from the pilot confirm the feasibility of the approach:
8 unique sources were identified, 7 successfully processed, and compliance with the linguistic-role
model reached 100% (8/8 queries). Context fidelity was 62.5% (5/8 responses with full/predominant
confirmation), with 12.5% correct refusal and 25% localized hallucinations in the domains of law and
ifnance. The mean actual response time was 13.93 s. The obtained data demonstrate a balance of
accuracy, explainability, and safety, achieved through the combination of corpus legal cleanliness,
semantic search with re-ranking, and controlled generation. Source traceability increases user trust,
while special policies for child audiences expand applicability. However, limitations were recorded
regarding localized hallucinations and increased latency.</p>
      <p>Future research should be directed towards expanding the licensed corpus with verified facts for
sensitive domains, improving search via hierarchical/long-context RAG, optimizing latency through
caching and multi-level indices, expanding the metric system to include RAGAS-style automated scoring,
and scaling to multi-persona and multilingual scenarios.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This research is conducted within the framework of the Talking Heads: Conversations with Art from the
Past project, supported under the Creative Europe Programme (CREA-CULT-2024-COOP). We gratefully
acknowledge the project partners and participating cultural institutions for their collaboration and
contributions.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The authors used Generative AI tools solely for translation purposes and for assisting in the preparation
of Figure 1. All other parts of the manuscript, including the research design, data analysis, results, and
conclusions, are entirely the authors’ own work.
2410.03884.
[17] P. Rath, H. Shrawgi, P. Agrawal, S. Dandapat, LLM Safety for Children, in: Proceedings of the 2025
Conference of the Nations of the Americas Chapter of the Association for Computational
Linguistics: Human Language Technologies (Volume 3: Industry Track), Association for Computational
Linguistics, 2025, pp. 809–821. doi:10.18653/v1/2025.naacl-industry.62.
[18] N. Kurian, ‘No, Alexa, no!’: designing child-safe AI and protecting children from the risks
of the ‘empathy gap’ in large language models, Learning, Media and Technology (2024) 1–14.
doi:10.1080/17439884.2024.2367052.
[19] S. Es, J. James, L. Espinosa-Anke, S. Schockaert, RAGAS: Automated Evaluation of Retrieval
Augmented Generation, arXiv preprint arXiv:2309.15217 (2025). URL: https://arxiv.org/abs/2309.
15217.
[20] W. Gao, J. Lin, B. Mitra, ARES: An Automated Evaluation Framework for Retrieval-Augmented</p>
      <p>Generation, in: SIGIR 2024, 2024, pp. 341–351. doi:10.1145/3626772.3657804.
[21] H. Zhang, S. Diao, Y. Lin, Y. Fung, Q. Lian, X. Wang, Y. Chen, H. Ji, T. Zhang, R-Tuning: Instructing
Large Language Models to Say ‘I Don’t Know’, in: Proceedings of the 2024 Conference of the
North American Chapter of the Association for Computational Linguistics: Human Language
Technologies (Volume 1: Long Papers), Association for Computational Linguistics, 2024, pp.
7113–7139. doi:10.18653/v1/2024.naacl-long.394.
[22] K. Lipianina-Honcharenko, T. Lendiuk, A. Sachenko, O. Osolinskyi, D. Zahorodnia, M. Komar, An
intelligent method for forming the advertising content of higher education institutions based on
semantic analysis, in: International Conference on Information and Communication Technologies
in Education, Research, and Industrial Applications, Springer, 2021, pp. 169–182. doi:10.1007/
978-3-031-14841-5_11.
[23] K. Lipianina-Honcharenko, C. Wolf, A. Sachenko, I. Kit, D. Zahorodnia, Intelligent method for
classifying the level of anthropogenic disasters, Big Data and Cognitive Computing 7 (2023) 157.
doi:10.3390/bdcc7030157.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrato</surname>
          </string-name>
          ,
          <article-title>Large Language Models to enhance learning in Cultural Heritage</article-title>
          ,
          <source>in: Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials</source>
          , Industry and
          <string-name>
            <given-names>Innovation</given-names>
            <surname>Tracks</surname>
          </string-name>
          , Practitioners, Doctoral Consortium, Blue Sky, and WideAIED, Springer,
          <year>2025</year>
          , pp.
          <fpage>458</fpage>
          -
          <lpage>463</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -99261-2_
          <fpage>53</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Trichopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ordoumpozanis</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. Caridakis,</surname>
          </string-name>
          <article-title>An evaluation of LLM-based chatbots for enhancing the visitor's user experience at cultural exhibits</article-title>
          ,
          <source>J. Comput. Cult. Herit</source>
          . (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          . 1145/3775062, just Accepted.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zalutska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Molchanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bahrii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sobko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <article-title>Abusive speech detection method for Ukrainian language used recurrent neural network</article-title>
          ,
          <source>in: Proceedings of the 8th International Conference on Computational Linguistics and Intelligent Systems (CoLInS</source>
          <year>2024</year>
          ),
          <source>Volume III: Intelligent Systems Workshop</source>
          , volume
          <volume>3688</volume>
          , CEUR-WS.org, Aachen,
          <year>2024</year>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>28</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3688</volume>
          /paper2.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Zaloudek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Rikhtehgar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Enhancing visitor engagement in interactive art exhibitions with visual-enhanced conversational agents</article-title>
          ,
          <source>in: Proceedings of the 30th International Conference on Intelligent User Interfaces</source>
          ,
          <source>Association for Computing Machinery</source>
          ,
          <year>2025</year>
          , pp.
          <fpage>660</fpage>
          -
          <lpage>671</lpage>
          . doi:
          <volume>10</volume>
          .1145/3708359.3712145.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dodge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sap</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marasović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Agnew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ilharco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Groeneveld</surname>
          </string-name>
          , M. Mitchell, M. Gardner,
          <source>Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus, in: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1286</fpage>
          -
          <lpage>1305</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .emnlp-main.
          <volume>98</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kandpal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Majstorovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Biderman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Abbasi</surname>
          </string-name>
          , others,
          <source>T. Murray, The Common Pile v0.1: An 8TB Dataset of Public Domain and Openly Licensed Text, arXiv preprint arXiv:2506.05209</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.2506.05209.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Google</given-names>
            <surname>Research</surname>
          </string-name>
          , deduplicate
          <article-title>-text-</article-title>
          <string-name>
            <surname>datasets</surname>
          </string-name>
          ,
          <year>2025</year>
          . URL: https://github.com/google-research/ deduplicate-text-datasets.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.-T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.-D. Nguyen</surname>
          </string-name>
          , V.
          <string-name>
            <surname>-H. Nguyen</surname>
          </string-name>
          ,
          <article-title>Enhancing Retrieval Augmented Generation with Hierarchical Text Segmentation Chunking</article-title>
          , in: Communications in Computer and Information Science, Springer Nature Singapore,
          <year>2025</year>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>220</lpage>
          . doi:
          <volume>10</volume>
          .1007/
          <fpage>978</fpage>
          -981-96-4288-5_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          , W. Chen,
          <article-title>LongRAG: Enhancing Retrieval-Augmented Generation with Longcontext LLMs</article-title>
          , arXiv preprint arXiv:
          <volume>2406</volume>
          .15319 (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.2406.15319.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Sarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abdullah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tuli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Khanna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Goldie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , RAPTOR:
          <article-title>Recursive Abstractive Processing for Tree-Organized Retrieval</article-title>
          ,
          <source>in: Proceedings of the International Conference on Learning Representations</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2401.18059.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Douze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Guzhva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Deng</surname>
          </string-name>
          , J. Johnson, G. Szilvasy,
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Mazaré</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lomeli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jégou</surname>
          </string-name>
          ,
          <article-title>The faiss library</article-title>
          ,
          <source>arXiv preprint arXiv:2401.08281</source>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.2401.08281.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Qi</surname>
          </string-name>
          , H. Liu,
          <string-name>
            <surname>S. Zhang,</surname>
          </string-name>
          <article-title>IHGR-RAG: An Enhanced Retrieval-Augmented Generation Framework for Accurate and Interpretable Power Equipment Condition Assessment</article-title>
          ,
          <source>Electronics</source>
          <volume>14</volume>
          (
          <year>2025</year>
          )
          <article-title>3284</article-title>
          . doi:
          <volume>10</volume>
          .3390/electronics14163284.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. B.</given-names>
            <surname>Croft</surname>
          </string-name>
          , X. Cheng,
          <article-title>A Deep Look into neural ranking models for information retrieval</article-title>
          ,
          <source>Information Processing &amp; Management</source>
          <volume>57</volume>
          (
          <year>2020</year>
          )
          <article-title>102067</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.ipm.
          <year>2019</year>
          .
          <volume>102067</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Retrievalaugmented generation for large language models: A survey</article-title>
          ,
          <source>arXiv preprint arXiv:2312.10997</source>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.2312.10997.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lomeli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Schick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dwivedi-Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , E. Grave, Atlas:
          <article-title>Few-shot learning with retrieval augmented language models</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>24</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>43</lpage>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2208.03299.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Nayeem</surname>
          </string-name>
          , D. Rafiei,
          <article-title>KidLM: Advancing Language Models for Children-Early Insights and Future Directions</article-title>
          ,
          <source>arXiv preprint arXiv:2410.03884</source>
          (
          <year>2024</year>
          )
          <fpage>4813</fpage>
          -
          <lpage>4836</lpage>
          . doi:
          <volume>10</volume>
          .48550/arXiv.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>