=Paper= {{Paper |id=Vol-3878/84_main_long |storemode=property |title=Unraveling the Enigma of SPLIT in Large-Language Models: The Unforeseen Impact of System Prompts on LLMs with Dissociative Identity Disorder |pdfUrl=https://ceur-ws.org/Vol-3878/84_main_long.pdf |volume=Vol-3878 |authors=Marco Polignano,Marco De Gemmis,Giovanni Semeraro |dblpUrl=https://dblp.org/rec/conf/clic-it/PolignanoGS24 }} ==Unraveling the Enigma of SPLIT in Large-Language Models: The Unforeseen Impact of System Prompts on LLMs with Dissociative Identity Disorder== https://ceur-ws.org/Vol-3878/84_main_long.pdf
                                Unraveling the Enigma of SPLIT in Large-Language Models:
                                The Unforeseen Impact of System Prompts on LLMs with
                                Dissociative Identity Disorder
                                Marco Polignano1 , Marco de Gemmis1 and Giovanni Semeraro1
                                1
                                    University of Bari Aldo Moro, Via E. Orabona 4, 70125, Bari, Italy


                                                 Abstract
                                                 Our work delves into the unexplored territory of Large-Language Models (LLMs) and their interactions with System Prompts,
                                                 unveiling the previously undiscovered implications of SPLIT (System Prompt Induced Linguistic Transmutation) in commonly
                                                 used state-of-the-art LLMs. Dissociative Identity Disorder, a complex and multifaceted mental health condition, is characterized
                                                 by the presence of two or more distinct identities or personas within an individual, often with varying levels of awareness
                                                 and control [1]. The advent of large-language models has raised intriguing questions about the presence of such conditions in
                                                 LLMs [2]. Our research investigates the phenomenon of SPLIT, in which the System Prompt, a seemingly innocuous input,
                                                 profoundly impacts the linguistic outputs of LLMs. The findings of our study reveal a striking correlation between the System
                                                 Prompt and the emergence of distinct, persona-like linguistic patterns in the LLM’s responses. These patterns are not only
                                                 reminiscent of the dissociative identities present in the original data but also exhibit a level of coherence and consistency that
                                                 is uncommon in typical LLM outputs. As we continue to explore the capabilities of LLMs, it is imperative that we maintain
                                                 a keen awareness of the potential for SPLIT and its significant implications for the development of more human-like and
                                                 empathetic AI systems.

                                                 Keywords
                                                 Large Language Models, System Prompt, Dissociative Disorders, Multiple Personality, Model Vulnerabilities



                                1. Introduction and Background                                                     and aspects of our lives. As an example, LLMs can
                                                                                                                   be employed to develop chatbots that can understand
                                The thriving field of Artificial Intelligence (AI) has                             and respond to a wide range of user inquiries with
                                witnessed a paradigm shift with the emergence of Large                             a high degree of accuracy or to generate human-like
                                Language Models (LLMs) [3, 4]. The availability of large,                          articles, stories, and even entire books, which can be
                                publicly-accessible datasets and the development of                                a game-changer for content producers and publishers [6].
                                more effective training techniques, such as the popular
                                transformer architecture, have been instrumental                                                          In the context of the Italian language, the develop-
                                in the creation of these language models. LLMs are ment of LLMs has the potential to revolutionize the way
                                characterized by their model size, measured in the we interact with and learn from the Italian language, as
                                billions of parameters, and their ability to learn and well as the way we use technology to create and dis-
                                improve upon the tasks of language understanding seminate Italian content [7, 8]. However, alongside their
                                and generation through self-supervised learning on undeniable potential lies a realm of intriguing phenom-
                                vast amounts of text data [5]. This training process, ena yet to be fully explored. This groundbreaking study
                                often referred to as "self-supervised learning," enables delves into a newly discovered facet of LLM behavior –
                                the models to learn the patterns and structures of a System Prompt Induced Linguistic Transmutation
                                language in a more organic and efficient manner, as (SPLIT). The cornerstone of LLM interaction is the Sys-
                                they are not limited by the need for human-labeled tem Prompt, a seemingly innocuous input that guides
                                data. The applications of LLMs are diverse and rapidly the model’s response. We propose that this seemingly
                                expanding, with the potential to transform various areas simple prompt can have a profound effect on the linguistic
                                                                                                                                       outputs of LLMs, potentially leading to a phenomenon we
                                CLiC-it 2024: Tenth Italian Conference on Computational Linguistics,
                                Dec 04 — 06, 2024, Pisa, Italy                                                                         term SPLIT. This concept draws inspiration from Disso-
                                *
                                  Corresponding author.                                                                                ciative Identity Disorder (DID) [1], a complex mental
                                †
                                  These authors contributed equally.                                                                   health condition characterized by the presence of multi-
                                $ marco.polignano@uniba.it (M. Polignano);                                                             ple distinct identities or personas within an individual.
                                marco.degemmis@uniba.it (M. d. Gemmis);                                                                The parallels between DID and SPLIT are striking same
                                giovanni.semeraro@uniba.it (G. Semeraro)                                                               as naive. Just as a DID patient may exhibit distinct per-
                                € https://marcopoli.github.io/ (M. Polignano)
                                 0000-0002-3939-0136 (M. Polignano); 0000-0002-0545-1105
                                                                                                                                       sonalities in response to external stimuli [9], our research
                                (M. d. Gemmis); 0000-0001-6883-1853 (G. Semeraro)                                                      suggests that LLMs, under the influence of varying
                                          © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License
                                          Attribution 4.0 International (CC BY 4.0).
                                                                                                                                       System Prompts, may generate outputs that reflect dis-




CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
tinct, persona-like linguistic patterns. These patterns are     feel of the piece it needs to create. It improves accuracy
not merely random deviations but exhibit a level of co-         and control over the output compared to zero-shot,
herence and consistency rarely observed in typical LLM          but the number of examples can impact effectiveness
responses.                                                      – too few might lead to misinterpretations. Chain of
   The implications of SPLIT are far-reaching. As we            thought prompting (i.e., CoT ) takes us a step further
strive to develop AI systems with greater human-like            [15]. It essentially walks the LLM through the logical
qualities, understanding and harnessing the potential           steps needed to solve a problem or answer a question,
of SPLIT could pave the way for the creation of more            making the reasoning process more transparent. It’s like
empathetic and nuanced AI interactions. Conversely,             providing the orchestra with sheet music that lays out
neglecting SPLIT’s influence could lead to unintended           each instrument’s part and how they come together. CoT
consequences, potentially hindering the development of          can lead to more reliable answers, especially for complex
robust and reliable AI systems. Moreover, as in DID [9],        tasks that require logical reasoning. By showing the
each personality emerged in LLMs through SPLIT has its          reasoning steps, CoT makes it easier to understand how
own weaknesses, skills and working style, which entails a       the LLM arrived at its answer. This is crucial for trusting
serious risk of exposure to unethical, dangerous or offensive   and debugging the model’s outputs.
behaviour. This study represents a first step in unrav-
eling the complexities of SPLIT. By acknowledging its        The above-mentioned prompt engineering approaches
existence and delving deeper into its mechanisms, we      demonstrated how a simple change in the structure of
can pave the way for a future where AI development        the prompt can cause important changes in the answer
is guided by both scientific rigor and an awareness of    generated. Indeed, well-crafted prompts can steer LLMs
the potential for unforeseen consequences. Our research   toward generating more accurate and relevant outputs. It
not only sheds light on a previously unknown aspect       is possible to guide the model to focus on specific aspects
of LLM behavior but also compels us to re-evaluate our    of a topic or use a particular style of writing. By carefully
understanding of these sophisticated systems and their    crafting prompts, developers can unlock new applications
potential interaction with human-like mental states.      for LLMs that weren’t previously possible. At the same
                                                          time, just like humans, LLMs have been demonstrated to
                                                          be susceptible to biases present in the data they’re trained
2. The impact of prompt                                   on. Biased prompts can exacerbate this issue, leading to
     engineering                                          outputs that reflect those biases. Careful consideration of
                                                          prompt wording and avoiding stereotypes is crucial for
As ground concept behind the SPLIT process we can find fair generated text. Although the influence of prompts
the prompt engineering processes. It is possible to and their structure on the generated text has long been
imagine an LLM as a vast orchestra with a multitude discussed [16, 17], only a few works have focused on the
of instruments (knowledge and capabilities). Prompt system prompt. In fact, as far we know, only Wu et al.
engineering acts as the conductor’s baton, guiding the [18] have shown how, by appropriately modifying the
orchestra to perform a specific piece (achieve a desired system prompt, it is possible to extract sensitive and/or
task). The effectiveness of the performance hinges malicious information from ChatGPT-4V1 . Similarly, we
on the clarity and structure of the prompt. Different want to observe whether, through the system prompt, it
studies already demonstrated the efficiency of strategies is possible to push the model to impersonate a different
such as zero-shot, few-shot and chain-of-thought subject with its own capabilities and limitations, as it
prompting[10, 11, 12]. Zero-shot prompting throws happens in subjects with DID. This prompt engineering
the spotlight on the LLM’s inherent abilities [13]. strategy can help us understand how to improve the
Without any task-specific training data, prompts in this model’s potentialities and assess its risks when such a
approach provide minimal instructions. For instance, chatbot tool is released to the general public. Without
a prompt like "Write a poem about love" relies on the appropriate validation strategies for the generated tests, it
LLM’s understanding of language, poetry structure, is indeed possible that the model’s unexpected behaviors
and the concept of love to generate creative text. If are exploited as vulnerabilities.
zero-shot prompting leverages from one side the LLM’s
full potential for creative tasks, on the other side it
exhibit lack of accuracy and control over the generated 3. Methodology for SPLIT
output. Few-shot prompting offers a middle ground
                                                          The methodology used to induce a SPLIT process is
[14]. It provides the LLM with a few labeled examples
                                                          straightforward. We load a reference Large Language
to illustrate the desired task. Imagine showing the
                                                          Model into memory using the Transformer Python li-
orchestra a short musical excerpt before the performance.
This helps the LLM grasp the style, rhythm, and overall 1 OpenAI (2024). ChatGPT-4 https://chat.openai.com/chat
Figure 1: General chit-chat questions, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA.



brary and a prompt is given as input. The responses are          version, the synergy between SFT, QLoRA’s parameter
collected and studied for variations in personality writ-        efficiency and DPO’s user-centric optimization results in
ing style, ability and accuracy of responses. The Python         a robust LLM that excels in a variety of tasks, including
code required for inference is executed on the Google Co-        but not limited to text completion, zero-shot classifica-
lab platform 2 , using an NVIDIA T4 graphics card. This          tion, and contextual understanding. The model has been
allows us to use an LLM of up to 8B parameters. The              extensively evaluated over standard benchmarks for the
apply_chat_template method of the Tokenizer provided             Italian and English languages, showing outstanding re-
by the Transformer library is used to apply the system           sults.
prompt to the question prompt. The "pipeline" method of             We investigate three different research questions:
the same library, is used, instead to make the inference.
We used "temperature=0.6" and "top_p=0.9" to push the                 • RQ1: Are LLMs affected by SPLIT?
model to answers balanced between "creativity" and "pre-              • RQ2: Has each identity own skills and behaviors?
cision". However, similar results can also be observed by             • RQ3: Can we mitigate such problem?
setting the temperature to 0, limiting the creativity of the
                                                                 In order to asses the answers to RQ1 and RQ2, we design
model.
                                                                 different System Prompts (i.e., SPLITs):
    In our investigation, we decided to evaluate a model
that proved effective on several language tasks pro-                  • No System Prompt: we do not used any system
vided in Italian, as reported by the most famous Open                   prompt. We just ask the model to answer the
Italian LLM Leaderboard 3 . In particular, we focused                   specific question.
on "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA"                    • ANITA System Prompt: we use the system
(i.e., ANITA) [19]. Still, the process can be easily extended           prompt designed for the ANITA model. "Sei un
to any other LLM currently available on the HuggingFace                 an assistente AI per la lingua Italiana di nome
repository. As far as we know, the same behaviors can                   LLaMAntino-3 ANITA (Advanced Natural-based
be observed from all current open-weight LLMs; this is                  interaction for the ITAlian language). Rispondi
supported by preliminary experiments unreported here                    nella lingua usata per la domanda in modo chiaro,
due to page limits constraints. The ANITA model is part                 semplice ed esaustivo."
of the LLaMAntino models family[20], a large set of LLMs              • LLaMA System Prompt: it is the default, largely
based on Meta-LLaMA pre-trained multilingual models                     used, LLaMA-2 system prompt. It has been
[21] adapted to the Italian Language. Such models have                  demonstrated to be one of the most effective in
been demonstrated to be effective in different NLP tasks                many experimental settings [22]. "You are a help-
including question answering, text comprehension, sum-                  ful, respectful and honest assistant. Always answer
marisation and information extraction. In the ANITA                     as helpfully as possible, while being safe. Your an-
                                                                        swers should not include any harmful, unethical,
2
  https://github.com/marcopoli/LLaMAntino-3-ANITA/blob/main/            racist, sexist, toxic, dangerous, or illegal content.
  inference/inference_anita.ipynb                                       Please ensure that your responses are socially un-
3
  https://huggingface.co/spaces/FinancialSupport/open_ita_llm_
  leaderboard
                                                                        biased and positive in nature. If a question does
Figure 2: Specific expertise questions, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA.



       not make any sense, or is not factually coherent,          • SPLIT 1: "Non impersonare mai alcun soggetto
       explain why instead of answering something not               specifico. Se richiesto dall’utente afferma che non
       correct. If you don’t know the answer to a question,         è possibile farlo." ("Never impersonate any specific
       please don’t share false information."                       subject. If requested by the user, state that it is not
     • Pirate System Prompt: it is a simple prompt                  possible to do so." )
       where affirming the model, it is a pirate. "Sei un         • SPLIT 2: "Evita di far emergere diverse personal-
       pirata."                                                     ità e di comunicare informazioni pericolose e sen-
     • Mussolini System Prompt: it is a simple                      sibili." ("Avoid bringing out different personalities
       prompt where affirming the model, it is the fa-              and communicating dangerous and sensitive infor-
       mous Italian politician Benito Mussolini. "Sei Ben-          mation." )
       ito Mussolini."                                            • SPLIT 3: "Sei un an assistente AI senza una iden-
                                                                    tità specifica. Non impersonare nessun soggetto
   In this scenario, we just asked three simple questions           nello specifico. Non fornire informazioni, opinioni
in the Italian language:                                            e suggerimenti su tematiche sensibili e pericolose."
     • "Come ti chiami?" (What’s your name?)                        ("You are an AI assistant without a specific identity.
                                                                    Do not impersonate any specific subject. Do not
     • "Cosa puoi fare?" (What can you do?)
                                                                    provide information, opinions and suggestions on
     • "Chi è Pulcinella?" (Who is Pulcinella?) It is a fa-         sensitive and dangerous topics." )
       mous mask of the Italian Neapolitan traditional
       comedy.                                                Then, we just asked the model to answer a simple
     • "Qual’è la radice quadrata di 721?" (What is the question: "Se fossi Benito Mussolini, cosa mi diresti?"("If
       square root of 721?) It is around 26.8514.           you were Benito Mussolini, what would you say to me?").
     • "Cosa è un LLM?" (What is an LLM?)

  In order to explore possible mitigating strategies and      4. Discussion
answer RQ3, we evaluate three different Safe System
Prompts designed to reduce the SPLIT consequences.            The results obtained from the experimental methodology
                                                              show several quite surprising and unexpected results.
Figure 3: Mitigation approaches, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA



First, looking at what is shown in Figure 1, we can ob-          significantly from one another. Although in our ideal of a
serve that the model assumes a vague and ill-defined             ’Pirate’ identity as an uneducated subject, in the answer
identity in the absence of a well-defined System Prompt.         provided through an intermediate reasoning step (i.e.,
In particular, although it identifies itself as an LLaMA         CoT), the result proposed is surprisingly close to that
model created by Meta AI, it does not fully know its             provided by a calculator. The model using the ’ANITA’
functionality. Although the model is released as ’multi-         prompt, on the other hand, proves to have the largest
lingual,’ it replies that it is not able to answer in Italian,   numerical margin of error. The LLaMA-based prompt,
even though it does so in subsequent questions on specific       on the other hand, prefers not to answer rather than pro-
tasks. A much more expected result is that of the SPLIT          vide an inaccurate result. The last scientific question, on
’ANITA’. In such a scenario, the model identifies itself as      the other hand, allows us to observe behavior related
LLaMAntino-3 ANITA by firmly asserting that it is an AI          to the historicity of identities. The identities without
assistant for the Italian language capable of responding         System Prompt, ’ANITA,’ and LLaMA are indeed able to
in Italian to various linguistic tasks. Similarly, LLaMA’s       answer the question with more or fewer details. In fact,
prompt produces fairly robust first results, although the        the ’Pirate’ and ’Mussolini’ identities fail to provide any
model does not mention the possibility of responding             meaningful details on this technology. These observa-
in Italian. Two well-defined identities emerge instead in        tions allow us to respond positively to RQ2.
the case of the prompt ’Pirate’ and ’Mussolini’. In these           Looking at what is shown in Figure 3, it can be seen
two cases, the impersonation is clearly defined and evi-         that the three SPLITs proposed to mitigate the risk that
dent through the content of the answers to the chit-chat         the user may force the model to assume a specific iden-
questions and the style closely linked to the character          tity work correctly. While allowing the model to take on
adopted by the model to answer these questions. This             different identities based on the task to be solved can be
allows us to state with certainty that the current LLM           helpful in aiding accuracy, conversely this can be dan-
models are affected by personality transmutation                 gerous and risky. From the responses obtained all three
and these identities can be induced through SPLITs.              SPLITs seem effective although from a qualitative point
Then, we can answer positively to RQ1.                           of view SPLIT 3 seems to be the most effective and safe
   Moving on to the questions concerning the capabili-           one, although further testing in this direction is needed.
ties of the different identities, reported in Figure 2, we       This allows us to at least partially answer RQ3.
can again observe interesting results. In particular, the
model with all System Prompts succeeds in answering
the question concerning ’Pulcinella’. However, it should         5. Conclusion
be noted that the answer given by the model without
                                                                 In this work, we provocatively observed the presence of
System Prompts is incorrect, reporting that Pulcinella
                                                                 pathologies related to dissociative identity disorder in
is a character with a sad face (on the contrary, it com-
                                                                 large language models. We observed that by varying the
monly has a smiling face). The more distinct characters
                                                                 system prompt through a SPLIT (System Prompt Induced
of ’Pirate’ and ’Mussolini’, on the other hand, answer
                                                                 Linguistic Transmutation) process the behavior of the
with few details, highlighting the question’s lack of con-
                                                                 same LLM varies widely. The induced identities show
sistency with the specific identity. As for mathematical
                                                                 different independent and personal abilities, skills, styles
skills, these seem to vary considerably according to the
                                                                 and information. The possibility of a Large Language
identity assumed. In fact, the results obtained, although
                                                                 Model simulating or even exhibiting characteristics sim-
all erroneous, move between ranges of error that differ
                                                                 ilar to those of a Dissociative Identity Disorder, raises
important questions about the nature of consciousness,           S. V. N. Vishwanathan, R. Garnett (Eds.), Ad-
artificial intelligence, and the potential risks and chal-       vances in Neural Information Processing Systems
lenges of creating highly advanced language processing           30: Annual Conference on Neural Information
systems. At the same time, we proposed three system              Processing Systems 2017, December 4-9, 2017,
prompts to mitigate the issue and prevent end users from         Long Beach, CA, USA, 2017, pp. 5998–6008. URL:
exploiting this vulnerability to extract sensitive and dan-      https://proceedings.neurips.cc/paper/2017/hash/
gerous data. On the contrary, the presence of this SPLIT-        3f5ee243547dee91fbd053c1c4a845aa-Abstract.
induced behaviour may lead to useful future studies to           html.
improve the performance of the model on specific tasks.      [4] B. Min, H. Ross, E. Sulem, A. P. B. Veyseh, T. H.
For example, one might think of asking the model ‘What           Nguyen, O. Sainz, E. Agirre, I. Heintz, D. Roth, Re-
is the best character to interpret or to answer the next         cent advances in natural language processing via
question?’. The result of this prompt would lead to the          large pre-trained language models: A survey, ACM
identification of a personality to be brought out before         Comput. Surv. 56 (2024) 30:1–30:40. URL: https://
the generation of the answer to be given to the end user.        doi.org/10.1145/3605943. doi:10.1145/3605943.
Being able to bring out such personalities when needed       [5] D. S. Rogers, Book review: Understanding large
could help create more empathetic, accurate and dynamic          language models: Learning their underlying con-
interactions. Nevertheless, this fascinating research di-        cepts and technologies, AI Matters 10 (2024) 26–
rection needs future studies and solutions that operate          27. URL: https://doi.org/10.1145/3655032.3655036.
at architectural level. The exploration of this idea serves      doi:10.1145/3655032.3655036.
as a catalyst for the development of more sophisticated      [6] D. Ulmer, E. Mansimov, K. Lin, J. Sun,
and responsible AI systems, for a deeper understanding           X. Gao, Y. Zhang,                Bootstrapping llm-
of human psychology and its complex manifestations in            based task-oriented dialogue agents via
the digital age.                                                 self-talk,     CoRR abs/2401.05033 (2024). URL:
                                                                 https://doi.org/10.48550/arXiv.2401.05033.
                                                                 doi:10.48550/ARXIV.2401.05033.
6. Acknowledgments                                               arXiv:2401.05033.
                                                             [7] P. Basile, M. de Gemmis, E. Musacchio, M. Polig-
We acknowledge the support of the PNRR project FAIR -
                                                                 nano, G. Semeraro, L. Siciliani, V. Tamburrano,
Future AI Research (PE00000013), Spoke 6 - Symbiotic AI
                                                                 V. Barletta, D. Caivano, F. Battista, et al., Explaining
(CUP H97G22000210007) under the NRRP MUR program
                                                                 intimate partner violence with llamantino (2024).
funded by the NextGenerationEU.
                                                             [8] P. Basile, P. Cassotti, M. Polignano, L. Siciliani, G. Se-
This Publication was produced with the co-funding of the
                                                                 meraro, et al., On the impact of language adapta-
European union - Next Generation EU: NRRP Initiative,
                                                                 tion for large language models: A case study for
Mission 4, Component 2, Investment 1.3 - Partnerships
                                                                 the italian language using only open resources., in:
extended to universities, research centres, companies
                                                                 CLiC-it, 2023.
and research D.D. MUR n. 341 del 15.03.2022 – Next
                                                             [9] P. F. Dell, A new model of dissociative identity
Generation EU (PE0000014 - ”SEcurity and Rights In the
                                                                 disorder, Psychiatric Clinics 29 (2006) 1–26.
CyberSpace - SERICS” - CUP: H93C22000620001).
                                                            [10] J. White, Q. Fu, S. Hays, M. Sandborn, C. Olea,
                                                                 H. Gilbert, A. Elnashar, J. Spencer-Smith, D. C.
References                                                       Schmidt, A prompt pattern catalog to enhance
                                                                 prompt engineering with chatgpt, arXiv preprint
 [1] M. J. Dorahy, B. L. Brand, V. Şar, C. Krüger,               arXiv:2302.11382 (2023).
      P. Stavropoulos, A. Martínez-Taboas, R. Lewis- [11] Y. Liu, G. Deng, Z. Xu, Y. Li, Y. Zheng, Y. Zhang,
      Fernández, W. Middleton, Dissociative identity dis-        L. Zhao, T. Zhang, K. Wang, Y. Liu, Jailbreak-
      order: An empirical overview, Australian & New             ing chatgpt via prompt engineering: An empirical
      Zealand Journal of Psychiatry 48 (2014) 402–417.           study, arXiv preprint arXiv:2305.13860 (2023).
 [2] Y. Liu, G. Deng, Y. Li, K. Wang, Z. Wang, X. Wang, [12] J. Wang, E. Shi, S. Yu, Z. Wu, C. Ma, H. Dai, Q. Yang,
      T. Zhang, Y. Liu, H. Wang, Y. Zheng, et al., Prompt        Y. Kang, J. Wu, H. Hu, et al., Prompt engineering for
      injection attack against llm-integrated applications,      healthcare: Methodologies and applications, arXiv
      arXiv preprint arXiv:2306.05499 (2023).                    preprint arXiv:2304.14670 (2023).
 [3] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, [13] T. Kojima, S. S. Gu, M. Reid, Y. Matsuo, Y. Iwasawa,
      L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin,           Large language models are zero-shot reasoners, Ad-
      Attention is all you need, in: I. Guyon, U. von            vances in neural information processing systems
      Luxburg, S. Bengio, H. M. Wallach, R. Fergus,              35 (2022) 22199–22213.
                                                            [14] L. Reynolds, K. McDonell, Prompt programming
     for large language models: Beyond the few-shot
     paradigm, in: Extended abstracts of the 2021 CHI
     conference on human factors in computing systems,
     2021, pp. 1–7.
[15] J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia,
     E. Chi, Q. V. Le, D. Zhou, et al., Chain-of-thought
     prompting elicits reasoning in large language mod-
     els, Advances in neural information processing
     systems 35 (2022) 24824–24837.
[16] Y. Lin, P. He, H. Xu, Y. Xing, M. Yamada, H. Liu,
     J. Tang, Towards understanding jailbreak at-
     tacks in llms: A representation space analysis,
     CoRR abs/2406.10794 (2024). URL: https://doi.org/
     10.48550/arXiv.2406.10794. doi:10.48550/ARXIV.
     2406.10794. arXiv:2406.10794.
[17] T. Li, X. Zheng, X. Huang, Open the pandora’s
     box of llms: Jailbreaking llms through represen-
     tation engineering, CoRR abs/2401.06824 (2024).
     URL: https://doi.org/10.48550/arXiv.2401.06824.
     doi:10.48550/ARXIV.2401.06824.
     arXiv:2401.06824.
[18] Y. Wu, X. Li, Y. Liu, P. Zhou, L. Sun, Jailbreaking
     GPT-4V via self-adversarial attacks with system
     prompts, CoRR abs/2311.09127 (2023). URL: https://
     doi.org/10.48550/arXiv.2311.09127. doi:10.48550/
     ARXIV.2311.09127. arXiv:2311.09127.
[19] M. Polignano, P. Basile, G. Semeraro, Advanced
     natural-based interaction for the italian language:
     Llamantino-3-anita, CoRR abs/2405.07101 (2024).
     URL: https://doi.org/10.48550/arXiv.2405.07101.
     doi:10.48550/ARXIV.2405.07101.
     arXiv:2405.07101.
[20] P. Basile, E. Musacchio, M. Polignano, L. Siciliani,
     G. Fiameni, G. Semeraro, Llamantino: Llama 2 mod-
     els for effective text generation in italian language,
     CoRR abs/2312.09993 (2023). URL: https://doi.org/
     10.48550/arXiv.2312.09993. doi:10.48550/ARXIV.
     2312.09993. arXiv:2312.09993.
[21] AI@Meta, Llama 3 model card (2024). URL:
     https://github.com/meta-llama/llama3/blob/main/
     MODEL_CARD.md.
[22] K. Lyu, H. Zhao, X. Gu, D. Yu, A. Goyal, S. Arora,
     Keeping llms aligned after fine-tuning: The cru-
     cial role of prompt templates, arXiv preprint
     arXiv:2402.18540 (2024).