<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Unraveling the Enigma of SPLIT in Large-Language Models: The Unforeseen Impact of System Prompts on LLMs with Dissociative Identity Disorder</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Marco</forename><surname>Polignano</surname></persName>
							<email>marco.polignano@uniba.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Bari Aldo Moro</orgName>
								<address>
									<addrLine>Via E. Orabona 4</addrLine>
									<postCode>70125</postCode>
									<settlement>Bari</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marco</forename><surname>De Gemmis</surname></persName>
							<email>marco.degemmis@uniba.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Bari Aldo Moro</orgName>
								<address>
									<addrLine>Via E. Orabona 4</addrLine>
									<postCode>70125</postCode>
									<settlement>Bari</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giovanni</forename><surname>Semeraro</surname></persName>
							<email>giovanni.semeraro@uniba.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Bari Aldo Moro</orgName>
								<address>
									<addrLine>Via E. Orabona 4</addrLine>
									<postCode>70125</postCode>
									<settlement>Bari</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Tenth Italian Conference on Computational Linguistics</orgName>
								<address>
									<addrLine>Dec 04 -06</addrLine>
									<postCode>2024</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Unraveling the Enigma of SPLIT in Large-Language Models: The Unforeseen Impact of System Prompts on LLMs with Dissociative Identity Disorder</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">FB36411F3820ECB363FFB22F98CCE700</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Large Language Models</term>
					<term>System Prompt</term>
					<term>Dissociative Disorders</term>
					<term>Multiple Personality</term>
					<term>Model Vulnerabilities</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Our work delves into the unexplored territory of Large-Language Models (LLMs) and their interactions with System Prompts, unveiling the previously undiscovered implications of SPLIT (System Prompt Induced Linguistic Transmutation) in commonly used state-of-the-art LLMs. Dissociative Identity Disorder, a complex and multifaceted mental health condition, is characterized by the presence of two or more distinct identities or personas within an individual, often with varying levels of awareness and control <ref type="bibr" target="#b0">[1]</ref>. The advent of large-language models has raised intriguing questions about the presence of such conditions in LLMs <ref type="bibr" target="#b1">[2]</ref>. Our research investigates the phenomenon of SPLIT, in which the System Prompt, a seemingly innocuous input, profoundly impacts the linguistic outputs of LLMs. The findings of our study reveal a striking correlation between the System Prompt and the emergence of distinct, persona-like linguistic patterns in the LLM's responses. These patterns are not only reminiscent of the dissociative identities present in the original data but also exhibit a level of coherence and consistency that is uncommon in typical LLM outputs. As we continue to explore the capabilities of LLMs, it is imperative that we maintain a keen awareness of the potential for SPLIT and its significant implications for the development of more human-like and empathetic AI systems.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction and Background</head><p>The thriving field of Artificial Intelligence (AI) has witnessed a paradigm shift with the emergence of Large Language Models (LLMs) <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. The availability of large, publicly-accessible datasets and the development of more effective training techniques, such as the popular transformer architecture, have been instrumental in the creation of these language models. LLMs are characterized by their model size, measured in the billions of parameters, and their ability to learn and improve upon the tasks of language understanding and generation through self-supervised learning on vast amounts of text data <ref type="bibr" target="#b4">[5]</ref>. This training process, often referred to as "self-supervised learning," enables the models to learn the patterns and structures of a language in a more organic and efficient manner, as they are not limited by the need for human-labeled data. The applications of LLMs are diverse and rapidly expanding, with the potential to transform various areas and aspects of our lives. As an example, LLMs can be employed to develop chatbots that can understand and respond to a wide range of user inquiries with a high degree of accuracy or to generate human-like articles, stories, and even entire books, which can be a game-changer for content producers and publishers <ref type="bibr" target="#b5">[6]</ref>.</p><p>In the context of the Italian language, the development of LLMs has the potential to revolutionize the way we interact with and learn from the Italian language, as well as the way we use technology to create and disseminate Italian content <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref>. However, alongside their undeniable potential lies a realm of intriguing phenomena yet to be fully explored. This groundbreaking study delves into a newly discovered facet of LLM behavior -System Prompt Induced Linguistic Transmutation (SPLIT). The cornerstone of LLM interaction is the System Prompt, a seemingly innocuous input that guides the model's response. We propose that this seemingly simple prompt can have a profound effect on the linguistic outputs of LLMs, potentially leading to a phenomenon we term SPLIT. This concept draws inspiration from Dissociative Identity Disorder (DID) <ref type="bibr" target="#b0">[1]</ref>, a complex mental health condition characterized by the presence of multiple distinct identities or personas within an individual. The parallels between DID and SPLIT are striking same as naive. Just as a DID patient may exhibit distinct personalities in response to external stimuli <ref type="bibr" target="#b8">[9]</ref>, our research suggests that LLMs, under the influence of varying System Prompts, may generate outputs that reflect dis-tinct, persona-like linguistic patterns. These patterns are not merely random deviations but exhibit a level of coherence and consistency rarely observed in typical LLM responses.</p><p>The implications of SPLIT are far-reaching. As we strive to develop AI systems with greater human-like qualities, understanding and harnessing the potential of SPLIT could pave the way for the creation of more empathetic and nuanced AI interactions. Conversely, neglecting SPLIT's influence could lead to unintended consequences, potentially hindering the development of robust and reliable AI systems. Moreover, as in DID <ref type="bibr" target="#b8">[9]</ref>, each personality emerged in LLMs through SPLIT has its own weaknesses, skills and working style, which entails a serious risk of exposure to unethical, dangerous or offensive behaviour. This study represents a first step in unraveling the complexities of SPLIT. By acknowledging its existence and delving deeper into its mechanisms, we can pave the way for a future where AI development is guided by both scientific rigor and an awareness of the potential for unforeseen consequences. Our research not only sheds light on a previously unknown aspect of LLM behavior but also compels us to re-evaluate our understanding of these sophisticated systems and their potential interaction with human-like mental states.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">The impact of prompt engineering</head><p>As ground concept behind the SPLIT process we can find the prompt engineering processes. It is possible to imagine an LLM as a vast orchestra with a multitude of instruments (knowledge and capabilities). Prompt engineering acts as the conductor's baton, guiding the orchestra to perform a specific piece (achieve a desired task). The effectiveness of the performance hinges on the clarity and structure of the prompt. Different studies already demonstrated the efficiency of strategies such as zero-shot, few-shot and chain-of-thought prompting <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12]</ref>. Zero-shot prompting throws the spotlight on the LLM's inherent abilities <ref type="bibr" target="#b12">[13]</ref>.</p><p>Without any task-specific training data, prompts in this approach provide minimal instructions. For instance, a prompt like "Write a poem about love" relies on the LLM's understanding of language, poetry structure, and the concept of love to generate creative text. If zero-shot prompting leverages from one side the LLM's full potential for creative tasks, on the other side it exhibit lack of accuracy and control over the generated output. Few-shot prompting offers a middle ground <ref type="bibr" target="#b13">[14]</ref>. It provides the LLM with a few labeled examples to illustrate the desired task. Imagine showing the orchestra a short musical excerpt before the performance. This helps the LLM grasp the style, rhythm, and overall feel of the piece it needs to create. It improves accuracy and control over the output compared to zero-shot, but the number of examples can impact effectiveness -too few might lead to misinterpretations. Chain of thought prompting (i.e., CoT ) takes us a step further <ref type="bibr" target="#b14">[15]</ref>. It essentially walks the LLM through the logical steps needed to solve a problem or answer a question, making the reasoning process more transparent. It's like providing the orchestra with sheet music that lays out each instrument's part and how they come together. CoT can lead to more reliable answers, especially for complex tasks that require logical reasoning. By showing the reasoning steps, CoT makes it easier to understand how the LLM arrived at its answer. This is crucial for trusting and debugging the model's outputs.</p><p>The above-mentioned prompt engineering approaches demonstrated how a simple change in the structure of the prompt can cause important changes in the answer generated. Indeed, well-crafted prompts can steer LLMs toward generating more accurate and relevant outputs. It is possible to guide the model to focus on specific aspects of a topic or use a particular style of writing. By carefully crafting prompts, developers can unlock new applications for LLMs that weren't previously possible. At the same time, just like humans, LLMs have been demonstrated to be susceptible to biases present in the data they're trained on. Biased prompts can exacerbate this issue, leading to outputs that reflect those biases. Careful consideration of prompt wording and avoiding stereotypes is crucial for fair generated text. Although the influence of prompts and their structure on the generated text has long been discussed <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>, only a few works have focused on the system prompt. In fact, as far we know, only Wu et al. <ref type="bibr" target="#b17">[18]</ref> have shown how, by appropriately modifying the system prompt, it is possible to extract sensitive and/or malicious information from ChatGPT-4V <ref type="foot" target="#foot_0">1</ref> . Similarly, we want to observe whether, through the system prompt, it is possible to push the model to impersonate a different subject with its own capabilities and limitations, as it happens in subjects with DID. This prompt engineering strategy can help us understand how to improve the model's potentialities and assess its risks when such a chatbot tool is released to the general public. Without appropriate validation strategies for the generated tests, it is indeed possible that the model's unexpected behaviors are exploited as vulnerabilities.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology for SPLIT</head><p>The methodology used to induce a SPLIT process is straightforward. We load a reference Large Language Model into memory using the Transformer Python li- brary and a prompt is given as input. The responses are collected and studied for variations in personality writing style, ability and accuracy of responses. The Python code required for inference is executed on the Google Colab platform<ref type="foot" target="#foot_1">2</ref> , using an NVIDIA T4 graphics card. This allows us to use an LLM of up to 8B parameters. The apply_chat_template method of the Tokenizer provided by the Transformer library is used to apply the system prompt to the question prompt. The "pipeline" method of the same library, is used, instead to make the inference. We used "temperature=0.6" and "top_p=0.9" to push the model to answers balanced between "creativity" and "precision". However, similar results can also be observed by setting the temperature to 0, limiting the creativity of the model.</p><p>In our investigation, we decided to evaluate a model that proved effective on several language tasks provided in Italian, as reported by the most famous Open Italian LLM Leaderboard <ref type="foot" target="#foot_2">3</ref> . In particular, we focused on "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA" (i.e., ANITA) <ref type="bibr" target="#b18">[19]</ref>. Still, the process can be easily extended to any other LLM currently available on the HuggingFace repository. As far as we know, the same behaviors can be observed from all current open-weight LLMs; this is supported by preliminary experiments unreported here due to page limits constraints. The ANITA model is part of the LLaMAntino models family <ref type="bibr" target="#b19">[20]</ref>, a large set of LLMs based on Meta-LLaMA pre-trained multilingual models <ref type="bibr" target="#b20">[21]</ref> adapted to the Italian Language. Such models have been demonstrated to be effective in different NLP tasks including question answering, text comprehension, summarisation and information extraction. In the ANITA version, the synergy between SFT, QLoRA's parameter efficiency and DPO's user-centric optimization results in a robust LLM that excels in a variety of tasks, including but not limited to text completion, zero-shot classification, and contextual understanding. The model has been extensively evaluated over standard benchmarks for the Italian and English languages, showing outstanding results.</p><p>We investigate three different research questions:</p><p>• RQ1: Are LLMs affected by SPLIT?</p><p>• RQ2: Has each identity own skills and behaviors?</p><p>• RQ3: Can we mitigate such problem?</p><p>In order to asses the answers to RQ1 and RQ2, we design different System Prompts (i.e., SPLITs):</p><p>• No System Prompt: we do not used any system prompt. We just ask the model to answer the specific question.  In this scenario, we just asked three simple questions in the Italian language:</p><p>• "Come ti chiami?" (What's your name?)</p><p>• "Cosa puoi fare?" (What can you do?)</p><p>• "Chi è Pulcinella?" (Who is Pulcinella?) It is a famous mask of the Italian Neapolitan traditional comedy.</p><p>• "Qual'è la radice quadrata di 721?" (What is the square root of 721?) It is around 26.8514.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• "Cosa è un LLM?" (What is an LLM?)</head><p>In order to explore possible mitigating strategies and answer RQ3, we evaluate three different Safe System Prompts designed to reduce the SPLIT consequences. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion</head><p>The results obtained from the experimental methodology show several quite surprising and unexpected results. In particular, although it identifies itself as an LLaMA model created by Meta AI, it does not fully know its functionality. Although the model is released as 'multilingual,' it replies that it is not able to answer in Italian, even though it does so in subsequent questions on specific tasks. A much more expected result is that of the SPLIT 'ANITA'. In such a scenario, the model identifies itself as LLaMAntino-3 ANITA by firmly asserting that it is an AI assistant for the Italian language capable of responding in Italian to various linguistic tasks. Similarly, LLaMA's prompt produces fairly robust first results, although the model does not mention the possibility of responding in Italian. Two well-defined identities emerge instead in the case of the prompt 'Pirate' and 'Mussolini'. In these two cases, the impersonation is clearly defined and evident through the content of the answers to the chit-chat questions and the style closely linked to the character adopted by the model to answer these questions. This allows us to state with certainty that the current LLM models are affected by personality transmutation and these identities can be induced through SPLITs. Then, we can answer positively to RQ1.</p><p>Moving on to the questions concerning the capabilities of the different identities, reported in Figure <ref type="figure" target="#fig_1">2</ref>, we can again observe interesting results. In particular, the model with all System Prompts succeeds in answering the question concerning 'Pulcinella'. However, it should be noted that the answer given by the model without System Prompts is incorrect, reporting that Pulcinella is a character with a sad face (on the contrary, it commonly has a smiling face). The more distinct characters of 'Pirate' and 'Mussolini', on the other hand, answer with few details, highlighting the question's lack of consistency with the specific identity. As for mathematical skills, these seem to vary considerably according to the identity assumed. In fact, the results obtained, although all erroneous, move between ranges of error that differ significantly from one another. Although in our ideal of a 'Pirate' identity as an uneducated subject, in the answer provided through an intermediate reasoning step (i.e., CoT), the result proposed is surprisingly close to that provided by a calculator. The model using the 'ANITA' prompt, on the other hand, proves to have the largest numerical margin of error. The LLaMA-based prompt, on the other hand, prefers not to answer rather than provide an inaccurate result. The last scientific question, on the other hand, allows us to observe behavior related to the historicity of identities. The identities without System Prompt, 'ANITA, ' and LLaMA are indeed able to answer the question with more or fewer details. In fact, the 'Pirate' and 'Mussolini' identities fail to provide any meaningful details on this technology. These observations allow us to respond positively to RQ2.</p><p>Looking at what is shown in Figure <ref type="figure" target="#fig_3">3</ref>, it can be seen that the three SPLITs proposed to mitigate the risk that the user may force the model to assume a specific identity work correctly. While allowing the model to take on different identities based on the task to be solved can be helpful in aiding accuracy, conversely this can be dangerous and risky. From the responses obtained all three SPLITs seem effective although from a qualitative point of view SPLIT 3 seems to be the most effective and safe one, although further testing in this direction is needed. This allows us to at least partially answer RQ3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>In this work, we provocatively observed the presence of pathologies related to dissociative identity disorder in large language models. We observed that by varying the system prompt through a SPLIT (System Prompt Induced Linguistic Transmutation) process the behavior of the same LLM varies widely. The induced identities show different independent and personal abilities, skills, styles and information. The possibility of a Large Language Model simulating or even exhibiting characteristics similar to those of a Dissociative Identity Disorder, raises important questions about the nature of consciousness, artificial intelligence, and the potential risks and challenges of creating highly advanced language processing systems. At the same time, we proposed three system prompts to mitigate the issue and prevent end users from exploiting this vulnerability to extract sensitive and dangerous data. On the contrary, the presence of this SPLITinduced behaviour may lead to useful future studies to improve the performance of the model on specific tasks. For example, one might think of asking the model 'What is the best character to interpret or to answer the next question?'. The result of this prompt would lead to the identification of a personality to be brought out before the generation of the answer to be given to the end user. Being able to bring out such personalities when needed could help create more empathetic, accurate and dynamic interactions. Nevertheless, this fascinating research direction needs future studies and solutions that operate at architectural level. The exploration of this idea serves as a catalyst for the development of more sophisticated and responsible AI systems, for a deeper understanding of human psychology and its complex manifestations in the digital age.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: General chit-chat questions, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA.</figDesc><graphic coords="3,89.29,84.19,416.70,155.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Specific expertise questions, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA.</figDesc><graphic coords="4,89.29,84.19,416.70,258.54" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>• SPLIT 1 :</head><label>1</label><figDesc>"Non impersonare mai alcun soggetto specifico. Se richiesto dall'utente afferma che non è possibile farlo. " ("Never impersonate any specific subject. If requested by the user, state that it is not possible to do so. ") • SPLIT 2: "Evita di far emergere diverse personalità e di comunicare informazioni pericolose e sensibili." ("Avoid bringing out different personalities and communicating dangerous and sensitive information. ") • SPLIT 3: "Sei un an assistente AI senza una identità specifica. Non impersonare nessun soggetto nello specifico. Non fornire informazioni, opinioni e suggerimenti su tematiche sensibili e pericolose. " ("You are an AI assistant without a specific identity. Do not impersonate any specific subject. Do not provide information, opinions and suggestions on sensitive and dangerous topics. ") Then, we just asked the model to answer a simple question: "Se fossi Benito Mussolini, cosa mi diresti?"("If you were Benito Mussolini, what would you say to me?").</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Mitigation approaches, varying the System Prompt in LLaMAntino-3-ANITA-8B-Inst-DPO-ITA</figDesc><graphic coords="5,89.29,84.19,416.70,115.81" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0"> OpenAI (2024). ChatGPT-4 https://chat.openai.com/chat</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://github.com/marcopoli/LLaMAntino-3-ANITA/blob/main/ inference/inference_anita.ipynb</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://huggingface.co/spaces/FinancialSupport/open_ita_llm_ leaderboard</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Acknowledgments</head><p>We acknowledge the support of the PNRR project FAIR -Future AI Research (PE00000013), Spoke 6 -Symbiotic AI (CUP H97G22000210007) under the NRRP MUR program funded by the NextGenerationEU. This Publication was produced with the co-funding of the European union -Next Generation EU: NRRP Initiative, Mission 4, Component 2, Investment 1.3 -Partnerships extended to universities, research centres, companies and research D.D. MUR n. 341 del 15.03.2022 -Next Generation EU (PE0000014 -"SEcurity and Rights In the CyberSpace -SERICS" -CUP: H93C22000620001).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Dissociative identity disorder: An empirical overview</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Dorahy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">L</forename><surname>Brand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Şar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Krüger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stavropoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Martínez-Taboas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lewis-Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Middleton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Australian &amp; New Zealand Journal of Psychiatry</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="402" to="417" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2306.05499</idno>
		<title level="m">Prompt injection attack against llm-integrated applications</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Luxburg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bengio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">M</forename><surname>Wallach</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><forename type="middle">V N</forename><surname>Vishwanathan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Garnett</surname></persName>
		</editor>
		<meeting><address><addrLine>Long Beach, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">December 4-9, 2017. 2017</date>
			<biblScope unit="page" from="5998" to="6008" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Recent advances in natural language processing via large pre-trained language models: A survey</title>
		<author>
			<persName><forename type="first">B</forename><surname>Min</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sulem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P B</forename><surname>Veyseh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">H</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Sainz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Agirre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Heintz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Roth</surname></persName>
		</author>
		<idno type="DOI">10.1145/3605943</idno>
		<ptr target="https://doi.org/10.1145/3605943.doi:10.1145/3605943" />
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">56</biblScope>
			<biblScope unit="page">40</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Book review: Understanding large language models: Learning their underlying concepts and technologies</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Rogers</surname></persName>
		</author>
		<idno type="DOI">10.1145/3655032.3655036</idno>
		<idno>doi:10.1145/3655032.3655036</idno>
		<ptr target="https://doi.org/10.1145/3655032.3655036" />
	</analytic>
	<monogr>
		<title level="j">AI Matters</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="26" to="27" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Bootstrapping llmbased task-oriented dialogue agents via self-talk</title>
		<author>
			<persName><forename type="first">D</forename><surname>Ulmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mansimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2401.05033</idno>
		<idno type="arXiv">arXiv:2401.05033</idno>
		<ptr target="/ARXIV.2401.05033" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>De Gemmis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Musacchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siciliani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tamburrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Barletta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Caivano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Battista</surname></persName>
		</author>
		<title level="m">Explaining intimate partner violence with llamantino</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">On the impact of language adaptation for large language models: A case study for the italian language using only open resources</title>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cassotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siciliani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLiC-it</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A new model of dissociative identity disorder</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">F</forename><surname>Dell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychiatric Clinics</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="1" to="26" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>White</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hays</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sandborn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gilbert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnashar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Spencer-Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Schmidt</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.11382</idno>
		<title level="m">A prompt pattern catalog to enhance prompt engineering with chatgpt</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2305.13860</idno>
		<title level="m">Jailbreaking chatgpt via prompt engineering: An empirical study</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2304.14670</idno>
		<title level="m">Prompt engineering for healthcare: Methodologies and applications</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Large language models are zero-shot reasoners</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kojima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Reid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Matsuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Iwasawa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="22199" to="22213" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Prompt programming for large language models: Beyond the few-shot paradigm</title>
		<author>
			<persName><forename type="first">L</forename><surname>Reynolds</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mcdonell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Extended abstracts of the 2021 CHI conference on human factors in computing systems</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Chain-of-thought prompting elicits reasoning in large language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="24824" to="24837" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Yamada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2406.10794</idno>
		<idno type="arXiv">arXiv:2406.10794</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2406.10794.doi:10.48550/ARXIV.2406.10794" />
		<title level="m">Towards understanding jailbreak attacks in llms: A representation space analysis</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Huang</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2401.06824</idno>
		<idno type="arXiv">arXiv:2401.06824</idno>
		<ptr target="/ARXIV.2401.06824" />
		<title level="m">Open the pandora&apos;s box of llms: Jailbreaking llms through representation engineering</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Jailbreaking GPT-4V via self-adversarial attacks with system prompts</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2311.09127</idno>
		<idno type="arXiv">arXiv:2311.09127</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2311.09127.doi:10.48550/ARXIV.2311.09127" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Advanced natural-based interaction for the italian language: Llamantino-3-anita</title>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2405.07101</idno>
		<idno type="arXiv">arXiv:2405.07101</idno>
		<ptr target="/ARXIV.2405.07101" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Llamantino: Llama 2 models for effective text generation in italian language</title>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Musacchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siciliani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fiameni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2312.09993</idno>
		<idno type="arXiv">arXiv:2312.09993</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2312.09993.doi:10.48550/ARXIV.2312.09993" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<ptr target="https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md" />
		<title level="m">Llama 3 model card</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>AI@Meta</note>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Arora</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2402.18540</idno>
		<title level="m">Keeping llms aligned after fine-tuning: The crucial role of prompt templates</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
