<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Separating Linguistic Competence from Factual Knowledge in Large Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jaime Collado-Montañez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science (University of Jaén)</institution>
          ,
          <addr-line>Campus Las Lagunillas, s/n, Jaén, 23071</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Large Language Models (LLMs) have demonstrated remarkable capabilities in language understanding and generation, driven by advancements in deep neural networks However, the current trend of developing increasingly larger models to enhance task competence comes at a significant cost, including a substantial carbon footprint with detrimental environmental consequences. Furthermore, these models often internalize vast amounts of factual knowledge, leading to issues such as hallucinations and the use of outdated information. This research explores the hypothesis that linguistic competence-the ability to understand and produce natural language-can be separated from memorized factual knowledge and other cognitive skills in neural networks. We propose the development of “Fundamental Language Models” (FLMs), smaller, more eficient models focused on language understanding and reasoning. These FLMs will leverage external sources and tools, using techniques like Retrieval Augmented Generation (RAG), to access up-to-date factual knowledge, thereby potentially mitigating both environmental impact and factual inaccuracies. Our main objective is to understand the functioning of Large Language Models as reasoning engines, with a special focus on language models for Spanish.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large Language Model</kwd>
        <kwd>Fundamental Language Model</kwd>
        <kwd>Hallucination</kwd>
        <kwd>Retrieval Augmented Generation</kwd>
        <kwd>Explainability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>carefully curated datasets that emphasize linguistic structure and reasoning. This research will
investigate the feasibility and efectiveness of these approaches, exploring the trade-ofs between model size,
linguistic competence, and the efective utilization of external tools. By focusing on building agents
with a clear separation of language processing and knowledge access, this work aims to contribute to
the development of more sustainable, reliable, and ethically sound artificial intelligence systems.</p>
      <p>The remainder of this paper is structured as follows: Section 2 provides a review of the relevant
literature on LLM emergent abilities and limitations; Section 3 presents the research hypothesis and
objectives. Section 4 details the methodology employed in this thesis and three experiments proposed,
and Section 5 concludes by outlining specific research elements for discussion.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background and Related Work</title>
      <p>
        Transformer models are language models pretrained to understand language structures by using
semisupervised learning with huge amounts of data. Encoder transformers such as BERT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] or RoBERTa [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
use Masked Language Modeling (MLM) mainly while decoder or generative transformers like LLaMa [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ],
Mistral [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], and GPT [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] are trained using Causal Language Modeling (CLM).
      </p>
      <p>
        According to the following general definition of emergence, as stated by the Nobel prize-winning
physicist Philip Anderson [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]: “Emergence is when quantitative changes in a system result in qualitative
changes in behavior”, the rapidly growing size of such models, especially the generative ones, into
what we call LLMs is allowing them to showcase new emergent abilities such as reasoning [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Along
with these properties, this semi-supervised pretraining technique allows LLMs to memorize lots of
factual data [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] that, in some cases, may lead to problems such as hallucinations [14] and outdated
answers when training data does not include the latest events and news. Hallucinations in LLMs refer to
instances where the model generates responses that are not factual or grounded in reality but rather are
inferred from patterns in the training data. These hallucinations can occur when the model synthesizes
information based on statistical correlations in the data rather than true understanding [15].
      </p>
      <p>In addition to that, the use of large corpora of texts from various sources in the generation of
pre-trained models results in the model capturing stereotypical patterns present in the texts. This
issue, known as bias detection, is related to explainability but focuses on the detection, evaluation, and
mitigation of gender, profession, origin, ethnicity, or religion stereotypes present in trained models [16].
The problem has become a topic of interest beyond the field of AI algorithm research and is known as
fairness [17] due to its ethical and legal implications.</p>
      <p>Additionally, although they seem powerful in terms of results and predictions, large language models
have their own limitations. The most significant is opacity or lack of transparency [ 18]. This means
that the logic and internal functioning of these models are hidden from the user, which is a serious
disadvantage because it prevents a human, whether expert or not, from verifying, interpreting, and
understanding the system’s reasoning and how decisions are made. In other words, any suficiently
complex system acts as a black box when it is easier to experiment with than to understand [19].</p>
      <p>The study of “foundational” language models can help address bias and improve explainability by
focusing on core linguistic competence, separate from stored factual knowledge. This approach aligns
with long-standing linguistic debates, particularly Chomsky’s distinction between internal (I-language)
and external (E-language) systems. As Grafi [ 20] notes, viewing language as an internal cognitive
system, rather than a socially embedded one, raises questions still relevant when interpreting LLM
behavior. Recent empirical work supports this conceptual separation. Miaschi et al. [21] demonstrate
that LLMs vary in their ability to generate sentences that follow explicit morpho-syntactic and
syntactic constraints, highlighting clear limitations in linguistic control that are independent of factual
recall. These findings underscore the value of disentangling linguistic competence from knowledge
storage—central to the design of FLMs.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Hypothesis and Objectives</title>
      <p>This research is guided by the central hypothesis that it is possible to develop Fundamental Language
Models (FLMs): Small Language Models (SLMs) that primarily encode linguistic competence rather
than extensive factual knowledge and other cognitive capabilities. We propose that these FLMs can
efectively function as reasoning engines when coupled with external knowledge sources and tools,
potentially leading to more eficient, reliable, and ethically sound AI systems.</p>
      <p>With this framework in mind, the main objective of this research is to investigate the functioning of
Large Language Models as fundamental language processing engines. To achieve this, the following
secondary objectives have been defined:
1. Study the internal encoding of linguistic structures and their role in language understanding,
independent of factual knowledge.
2. Decompose the capabilities of current language models to diferentiate between core linguistic
skills and those reliant on internalized knowledge or reasoning.
3. Develop and evaluate methods for enhancing language model capabilities by integrating external
knowledge sources and tools for tasks beyond pure language processing.
4. Improve the explainability of AI task resolution by clearly separating the linguistic processing
stage from the contributions of external modules and retrieved information.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>The following methodology aims to validate the hypothesis that linguistic competence in LLMs can be
separated from other cognitive abilities, enabling the development of FLMs. The methodology comprises
the following key stages:
1. Literature review and initial study: A comprehensive review of existing literature will be conducted
to establish a strong foundation in current LLM techniques and advancements. This will involve
identifying key resources, including publications in leading scientific forums such as AAAI,
NeurIPS, and ACL, as well as relevant information from reference bulletins like PapersWithCode
and The Batch.
2. Experimental design and evaluation: Rigorous experimental setups will be defined for each
research objective. This includes specifying appropriate datasets and evaluation metrics. To
ensure robust evaluation and benchmarking, we will actively participate in evaluation forums
such as CLEF, SemEval, and IberLEF.
3. Study of internal encoding of knowledge: A series of experiments will be performed to analyze
the internal representations of LLMs. This will involve comparing representations across diferent
models to identify common patterns and structures related to linguistic competence encoding
and other capabilities.
4. Decomposition of language model capabilities: This stage focuses on isolating and evaluating
individual linguistic skills. Specific tasks and benchmarks will be designed or adapted to target
lexical, grammatical, and semantic competencies. Controlled experiments will be conducted
to assess model performance on these targeted tasks, allowing for a detailed analysis of each
competency.
5. Enhancement through external tools: Methods for efectively connecting LLMs with external
tools and knowledge bases will be developed and evaluated. Retrieval Augmented Generation
(RAG) will be a key technique explored. Experiments will compare the performance of enhanced
models against baseline models to quantify the benefits of such tools.
6. Dissemination of findings: Research findings will be prepared and disseminated through
publications in high-impact journals and presentations at international conferences.
4.1. Proposed experiments
This section outlines three key experiments currently in development. These experiments are designed to:
1) provide evidence supporting the FLM concept; 2) enable rigorous evaluation of linguistic competence,
a core feature of FLMs; and 3) demonstrate the potential of smaller models to perform complex tasks
when equipped with appropriate external tools and suficient linguistic competence.</p>
      <p>Study of internal encoding of knowledge: This experiment aims to analyze the scaling behavior
of linguistic competence in LLMs relative to reasoning and factual recall. Specialized benchmarks will
be used to assess each competency:
• Linguistic competence: Assessed as a composite of lexical, grammatical, and semantic
competencies. The hypothesis is that this competence will saturate at smaller model sizes compared to
reasoning and factual recall.
• External factual knowledge: Evaluated using question-answering tasks based on provided text
chunks. These tasks require reasoning and inference over the given factual information.
• Internal factual knowledge: Evaluated using question-answering tasks without context. Success
in these tasks relies on the model’s memorized factual knowledge.</p>
      <p>Lexical competence evaluation: Current benchmarks often lack the ability to efectively evaluate
LLMs’ lexical competence across diverse languages and specialized domains. To address this gap, we
are developing a novel, automated method for evaluating lexical competence in a multilingual and
domain-specific manner. This method will be used to assess the preservation of lexical competence in
FLMs as other capabilities are reduced.</p>
      <p>Enhancement through external tools: This experiment explores the use of external reasoning
tools to enhance the performance of small language models (SLMs) on logic-based benchmarks, such
as ZebraLogic [22]. The objective is to determine whether reasoning capabilities, similar to factual
knowledge in RAG systems, can be efectively delegated to external tools, or whether they represent a
more fundamental capability that FLMs must retain internally.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Research elements proposed for discussion</title>
      <p>The following key discussion points are proposed to facilitate a comprehensive exploration of the
potential benefits, limitations, and broader implications of the FLM paradigm within AI systems. These
points directly relate to the central hypothesis and objectives of this research:
• To what extent does separating linguistic competence from the storage of factual
knowledge mitigate the generation of hallucinations and enhance the reliability of information
generated by LLMs? This point prompts a critical evaluation of whether the FLM approach
efectively reduces the occurrence of inaccurate or speculative content arising from outdated or
erroneous internalized information.
• How efectively can external tools and retrieval mechanisms, such as Retrieval
Augmented Generation (RAG), provide FLMs with accurate, up-to-date information and
empower them to perform complex, knowledge-intensive tasks? This discussion will
center on the capabilities and limitations of RAG and other external knowledge integration
methods in supporting FLMs’ access to and utilization of current and relevant information.
• How does the separation of linguistic competence from other cognitive skills impact
the transparency of FLMs’ processing of information and task execution? This inquiry
explores whether separating linguistic competence from factual knowledge enhances the model’s
ability to explain its reasoning process transparently, thereby improving trust and interpretability
in AI-driven decision-making.
• To what extent can FLMs maintain comprehensive language understanding and
generation capabilities when completely decoupled from internalized knowledge, relying
solely on external resources? This discussion will critically assess the potential trade-ofs
between knowledge externalization and the preservation of essential linguistic competencies
required for efective performance in various AI applications.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work has been funded by the scholarship (FPI-PRE2022-105603) from the Ministry of Science,
Innovation and Universities of the Spanish Government. I am grateful to my thesis supervisors Arturo
Montejo-Ráez and L. Alfonso Ureña-López for their guidance and help during the work done up to
now.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author used ChatGPT and Grammarly in order to: Grammar
and spelling check, translate and reword. After using this tools, the author reviewed and edited the
content as needed and take full responsibility for the publication’s content.
[14] L. Huang, W. Yu, W. Ma, W. Zhong, Z. Feng, H. Wang, Q. Chen, W. Peng, X. Feng, B. Qin, T. Liu,
A survey on hallucination in large language models: Principles, taxonomy, challenges, and open
questions, 2023. arXiv:2311.05232.
[15] W. Wang, B. Haddow, A. Birch, W. Peng, Assessing factual reliability of large language model
knowledge, in: K. Duh, H. Gomez, S. Bethard (Eds.), Proceedings of the 2024 Conference of the
North American Chapter of the Association for Computational Linguistics: Human Language
Technologies (Volume 1: Long Papers), Association for Computational Linguistics, Mexico City,
Mexico, 2024, pp. 805–819. URL: https://aclanthology.org/2024.naacl-long.46. doi:10.18653/v1/
2024.naacl-long.46.
[16] I. Garrido-Muñoz , A. Montejo-Ráez , F. Martínez-Santiago , L. A. Ureña-López , A survey on
bias in deep nlp, Applied Sciences 11 (2021). URL: https://www.mdpi.com/2076-3417/11/7/3184.
doi:10.3390/app11073184.
[17] P. Hacker, Teaching fairness to artificial intelligence: existing and novel strategies against
algorithmic discrimination under eu law, Common market law review 55 (2018).
[18] M. Du, N. Liu, X. Hu, Techniques for interpretable machine learning, Communications of the</p>
      <p>ACM 63 (2019) 68–77.
[19] D. Golovin, B. Solnik, S. Moitra, G. Kochanski, J. Karro, D. Sculley, Google vizier: A service for
black-box optimization, in: Proceedings of the 23rd ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining, KDD ’17, Association for Computing Machinery, New
York, NY, USA, 2017, p. 1487–1495. URL: https://doi.org/10.1145/3097983.3098043. doi:10.1145/
3097983.3098043.
[20] G. Grafi, Between linguistics and philosophy of language: The debate on chomsky’s notion of
“knowledge of language”, Cahiers du Centre de Linguistique et des Sciences du Langage (2018)
39–58.
[21] A. Miaschi, F. Dell’Orletta, G. Venturi, Evaluating large language models via linguistic profiling,
in: Y. Al-Onaizan, M. Bansal, Y.-N. Chen (Eds.), Proceedings of the 2024 Conference on Empirical
Methods in Natural Language Processing, Association for Computational Linguistics, Miami,
Florida, USA, 2024, pp. 2835–2848. URL: https://aclanthology.org/2024.emnlp-main.166/. doi:10.
18653/v1/2024.emnlp-main.166.
[22] B. Y. Lin, R. L. Bras, K. Richardson, A. Sabharwal, R. Poovendran, P. Clark, Y. Choi, Zebralogic:
On the scaling limits of llms for logical reasoning, 2025. URL: https://arxiv.org/abs/2502.01100.
arXiv:2502.01100.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1] OpenAI, Gpt-4
          <source>technical report</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2303.08774. arXiv:
          <volume>2303</volume>
          .
          <fpage>08774</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Grattafiori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jauhri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kadian</surname>
          </string-name>
          ,
          <source>The llama 3 herd of models</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2407.21783. arXiv:
          <volume>2407</volume>
          .
          <fpage>21783</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Rillig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ågerstrand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Gould</surname>
          </string-name>
          , U. Sauerland,
          <article-title>Risks and benefits of large language models for the environment</article-title>
          ,
          <source>Environmental science &amp; technology 57</source>
          (
          <year>2023</year>
          )
          <fpage>3464</fpage>
          -
          <lpage>3466</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          , W. Ma,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Qin</surname>
          </string-name>
          , T. Liu,
          <article-title>A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions</article-title>
          ,
          <source>ACM Trans. Inf. Syst</source>
          .
          <volume>43</volume>
          (
          <year>2025</year>
          ). URL: https://doi.org/10.1145/3703155. doi:
          <volume>10</volume>
          .1145/ 3703155.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I. O.</given-names>
            <surname>Gallegos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Barrow</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Tanjim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Dernoncourt</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>N. K.</given-names>
          </string-name>
          <string-name>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <article-title>Bias and fairness in large language models: A survey</article-title>
          ,
          <source>Computational Linguistics</source>
          <volume>50</volume>
          (
          <year>2024</year>
          )
          <fpage>1097</fpage>
          -
          <lpage>1179</lpage>
          . URL: https://doi.org/10.1162/coli_a_00524. doi:
          <volume>10</volume>
          .1162/coli_a_
          <fpage>00524</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample,
          <article-title>Llama: Open and eficient foundation language models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A. Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sablayrolles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bamford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Chaplot</surname>
          </string-name>
          , D. de las Casas,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bressand</surname>
          </string-name>
          , G. Lengyel,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lample</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Saulnier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Lavaud</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Stock</surname>
            ,
            <given-names>T. L.</given-names>
          </string-name>
          <string-name>
            <surname>Scao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>W. E.</given-names>
          </string-name>
          <string-name>
            <surname>Sayed</surname>
          </string-name>
          , Mistral 7b,
          <year>2023</year>
          . arXiv:
          <volume>2310</volume>
          .
          <fpage>06825</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Narasimhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Salimans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Sutskever</surname>
          </string-name>
          , et al.,
          <article-title>Improving language understanding by generative pre-training (</article-title>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <article-title>More is diferent: Broken symmetry and the nature of the hierarchical structure of science</article-title>
          .,
          <source>Science</source>
          <volume>177</volume>
          (
          <year>1972</year>
          )
          <fpage>393</fpage>
          -
          <lpage>396</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bommasani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zoph</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Borgeaud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yogatama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Metzler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hashimoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fedus</surname>
          </string-name>
          ,
          <source>Emergent abilities of large language models</source>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2206</volume>
          .
          <fpage>07682</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Seo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.-S.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Seo</surname>
          </string-name>
          ,
          <article-title>How do large language models acquire factual knowledge during pretraining?</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2406.11813. arXiv:
          <volume>2406</volume>
          .
          <fpage>11813</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>