<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Unipa-GPT: a framework to assess open-source alternatives to Chat-GPT for Italian chat-bots</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Irene</forename><surname>Siragusa</surname></persName>
							<email>irene.siragusa02@unipa.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Engineering</orgName>
								<orgName type="institution">University of Palermo</orgName>
								<address>
									<postCode>90128</postCode>
									<settlement>Palermo, Sicily</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">IT University of Copenhagen</orgName>
								<address>
									<addrLine>København S</addrLine>
									<postCode>2300</postCode>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Roberto</forename><surname>Pirrone</surname></persName>
							<email>roberto.pirrone@unipa.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Engineering</orgName>
								<orgName type="institution">University of Palermo</orgName>
								<address>
									<postCode>90128</postCode>
									<settlement>Palermo, Sicily</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">Tenth Italian Conference on Computational Linguistics</orgName>
								<address>
									<addrLine>Dec 04 -06</addrLine>
									<postCode>2024</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Unipa-GPT: a framework to assess open-source alternatives to Chat-GPT for Italian chat-bots</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">AAF596CF9AA73BB3568AEA91C6E24D66</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>RAG</term>
					<term>ChatGPT</term>
					<term>LLM</term>
					<term>Embedding</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper illustrates the implementation of Open Unipa-GPT, an open-source version of the Unipa-GPT chat-bot that leverages open-source Large Language Models for embeddings and text generation. The system relies on a Retrieval Augmented Generation approach, thus mitigating hallucination errors in the generation phase. A detailed comparison between different models is reported to illustrate their performance as regards embedding generation, retrieval, and text generation. In the last case, models were tested in a simple inference setup after a fine-tuning procedure. Experiments demonstrate that an open-source LLMs can be efficiently used for embedding generation, but none of the models does reach the performances obtained by closed models, such as gpt-3.5-turbo in generating answers. Corpora and code are available on GitHub 1</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The increasing development of bigger and bigger Large Language Models (LLM), reaching 70B parameters as for Meta LLMs (Llama 2 <ref type="bibr" target="#b0">[1]</ref> and Llama 3 <ref type="bibr" target="#b1">[2]</ref>) and more as for OpenAI ones (GPT-3 <ref type="bibr" target="#b2">[3]</ref> and GPT-4 [4] 1 ), requires a significant computational resources for training, fine-tuning or inference. OpenAI models are accessible only upon payment via OpenAI API and cannot be downloaded in any way, while the open-source models by Meta are available also in the 8B and 13B parameters versions, and they can either be fine-tuned via Parameter-Efficient Fine-Tuning techniques (PEFT) <ref type="bibr" target="#b4">[5]</ref> such as LoRA <ref type="bibr" target="#b5">[6]</ref>, or they can make direct inference using a 8-bit quantization <ref type="bibr" target="#b6">[7]</ref> keeping the computational resources relatively small.</p><p>The availability of open-source small-size LLMs is crucial for developing Natural Language Process (NLP) applications that leverage a fine-tuning procedure over a specific domain or language, as for Anita <ref type="bibr" target="#b7">[8]</ref>, an Italian 8B adaptation of <ref type="bibr">Llama 3.</ref> Nevertheless, GPT and Llama models cannot be considered as truly open-source since their training data set is not available and, as for GPT models, and also their actual architecture is not accessible. Minerva <ref type="bibr" target="#b8">[9]</ref> model, on the other side, is an Italian and English LLM whose architecture, weights, and training data are accessible, but it</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related works</head><p>The increasing interest in developing Language Models (LM) for the Italian language, starts when BERT <ref type="bibr" target="#b11">[12]</ref> was first released and adapted models, such as AlBERTo <ref type="bibr" target="#b12">[13]</ref> were developed. After ChatGPT was made public <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>, an increasing interest in developing and using LLMs, and in generative AI based on decoder-only model, was crucial, also for the Italian NLP community, thus leading to the development of foundational models based on Llama 2 <ref type="bibr" target="#b0">[1]</ref> and Llama 3 <ref type="bibr" target="#b1">[2]</ref>. Among those models, LLaMantino (chat version) <ref type="bibr" target="#b13">[14]</ref> and Fauno <ref type="bibr" target="#b14">[15]</ref>, are based on Llama 2 fine-tuned for chat purposes, while Camoscio <ref type="bibr" target="#b15">[16]</ref> and Anita <ref type="bibr" target="#b7">[8]</ref> are a fine-tuned Italian version of the instruct version of Llama 2 and Llama 3, respectively. RAG is used in developing chat-bots which are grounded in various domains where the models need to be deeply guided in generation to avoid hallucination in their answers. Various examples can be found in the educational domain as for AI4LA <ref type="bibr" target="#b16">[17]</ref>, an assistant to students with Specific Learning Disorders (SLDs) like Dyslexia, Dysorthographia, and Dyscalculia, or as assistant providing information about restaurant industry <ref type="bibr" target="#b17">[18]</ref> or as chat-bot for Frequently Asked Questions (FAQ) <ref type="bibr" target="#b18">[19]</ref>. Also chat-bots for the Italian language were implemented for real-wold applications, namely as assistant for Italian Funding Application <ref type="bibr" target="#b19">[20]</ref>, or in the medical domain <ref type="bibr" target="#b20">[21]</ref> or in industrial context <ref type="bibr" target="#b21">[22]</ref>. The aforementioned works share the same architecture with the one we used to implement our model. In contrast with them, we decided to stress capabilities of open-source LLMs and do not rely on GPT-based models, that are used as baseline reference for text generation (gpt-3.5-turbo) and as an external judge to evaluate performances of the other models (gpt-4.5-turbo).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">System architecture</head><p>Open Unipa-GPT relies on two main components as it is shown in Figure <ref type="figure" target="#fig_0">1</ref> that is the Retriever and the Generator. In the following, the two components are detailed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Retriever</head><p>The Retriever is made up of a vector database built using the LangChain framework<ref type="foot" target="#foot_0">2</ref> , which makes use of the Facebook AI Similarity Search (FAISS) library <ref type="bibr" target="#b22">[23]</ref>. The vector database is filled with the documents belonging to the unipa-corpus (Appendix A), that are divided into 1K token chunks with an overlap of 50 tokens. Split documents are then processed by a LLM (the Embedding LLM) to generate the corresponding embedding, and store them in the vector database. Different LLMs were used for embedding generation: we selected the best models according to the Massive Text Embedding Benchmark (MTEB) <ref type="bibr" target="#b23">[24]</ref> for Information Retrieval<ref type="foot" target="#foot_1">3</ref> . We selected only models that explicitly state that they were trained and tested also with Italian data. In the end, we selected the following models: BGE-M3 (BGE) <ref type="bibr" target="#b24">[25]</ref>, E5-mistral-7b-instruct (E5-mistral) <ref type="bibr" target="#b25">[26]</ref>, sentence-bert-baseitalian-xxl-uncased 4 (BERT-it) and Multilingual-E5-large-instruct (m-E5) <ref type="bibr" target="#b26">[27]</ref> .</p><p>A vector database was built for each model, and their corresponding embedding spaces were compared to each other and with text-embedding-ada-002, the embedding model from OpenAI, to asses their retrieval performances (Section 5).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Generator</head><p>The Generator uses the following Italian isntruction prompt to answer to user questions: Both the question and the related relevant context are passed as input to the model, along with the prompt. As regards the Generator LLM, we used Transformer-based models <ref type="bibr" target="#b27">[28]</ref>. We choose not to use LLMs based on Llama 2 and deeply focused our work towards the most recent models, covering both Llama-and Mistral-based architectures. In particular, Llama-3-8B-instruct <ref type="bibr" target="#b1">[2]</ref> was used along with its adapted version for Italian, Anita-8B <ref type="bibr" target="#b7">[8]</ref>, and Minerva-3B <ref type="bibr" target="#b8">[9]</ref>, which is a Mistral-based architecture <ref type="bibr" target="#b28">[29]</ref>. All the generation LLMs were evaluated both in their base version and in the instruction-tuned one. The last ones were obtained via a three-epochs fine-tuning procedure with the Alpaca-LoRA <ref type="bibr" target="#b5">[6]</ref> strategy testing the Alpaca-LoRA hyper-parameters <ref type="foot" target="#foot_3">5</ref> for both 20 and 50 epochs. In the generation phase, models were asked to output at most 256 tokens. We manually generated a small set of Question-Answer (QA) pairs for evaluation starting from the real questions issued by the public during the 2023 SHARPER European Researchers' Night where Unipa-GPT was demonstrated. The procedure for building these QA pairs is reported in Section 4. We developed the entire system on a server with 2 Intel(R) Xeon(R) 6248R CPUs, 384 GB RAM, and two 48 GB NVIDIA RTX 6000 Ada Generation GPUs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">The data set</head><p>The Italian documents data set built for Unipa-GPT is called unipa-corpus <ref type="bibr" target="#b9">[10]</ref>, and it has been generated from scraping either HTML pages or PDF documents that are publicly available on the website of the University of Palermo, and it includes information about all the available Bachelor/Master degree courses in the academic year 2023/2024 along with practical information for future students, e.g. how to pay taxes, the enrollment procedure, and the related deadlines. Starting from this data set, a QA data set was created with a semisupervised procedure to allow instruction-tuning over general-purpose LLMs. Further information about the unipa-corpus is reported in Appendix A.</p><p>As already mentioned The original Unipa-GPT was available for public unsupervised QA during the European Researchers' Night in 2023, where a total of 165 questions was collected, along with feedback of users. On average, an interaction with the chat-bot was two questions long, and we collected qualitative evaluation of the user experience through a suitable questionnaire people were requested to fill on line just after having chatted with Unipa-GPT. Questionnaires were further analyzed, and resulted in a general positive evaluation of the system's performances by the majority of the users, which were mostly University students.</p><p>To generate the golden QA pairs used to assess the different performances of each generator LLM, we devised six typologies by the direct inspection of collected questions. Particularly we groupte questions in Generic Information, Courses' Information, Other University-related, Services and Structures, Taxes and Scholarships, University Environment, and Off-topic. Next, we picked one question per typology, discarding the Off-topic ones, and a golden answer was manually built for each of them by leveraging the actual relevant documents contained in the corpus, thus marking them as golden documents. Note that if an answer can be elicited by multiple documents, all of them have been marked as golden. The detailed list of the Italian QA pairs is reported in Appendix B in Table <ref type="table" target="#tab_3">4</ref>, while the English version is reported in Table <ref type="table" target="#tab_4">5</ref>. Note that the English version is reported here for full readability purposes, while only Italian data were used for evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experimental results</head><p>The proposed model is intended to work in an open QA context, where correct answers are not known, thus, after a previous phase of qualitative evaluation <ref type="bibr" target="#b9">[10]</ref> as in <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22]</ref>, we opted for a quantitative analysis, relying on the small QA data set described in Section 4 to evaluate the performances against a set of golden labels in terms of both retrieval and answering capabilities <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b17">18]</ref>.</p><p>For each QA test pair, we retrieved the four most relevant documents from each vector database related to one of the open Embedding LLMs under investiga- tion. Then we scored the retrieved documents in terms of their context relevancy with respect to the provided question using the RAGAS framework <ref type="bibr" target="#b30">[31]</ref> that exploits gpt-4-turbo for the evaluation. Results are reported in Table <ref type="table" target="#tab_1">1</ref>, and they include also the performances of the original vector database using OpenAI embeddings (text-embedding-ada-002, referred as open-ai-ada).</p><p>The overall scores are not so high, and also the highest relevancy do not always correspond to the golden document used for generating the corresponding answer. In Table <ref type="table" target="#tab_1">1</ref> the underlined values are the ones associated with golden documents, while the bold ones are the highest RAGAS values. A model is considered to perform correctly if the highest context relevancy score is assigned to one of the golden documents. This evaluation procedure led to select E5-mistral as the best performing Embeddings LLM among the ones we investigated. Superior performances of E5-mistral are also confirmed by a deep analysis on the embeddings space by means of two different clustering procedures. We clustered the embeddings generated by each LLM starting from the documents belonging to both sections Educational Offer and Future Students of the UniPA website. The firs group of documents is the list of all the available courses at the University, while the second group contains useful information for future students who want to enroll in a degree course. We clustered the embedding spaces according to the either the course degree typology (bachelor/master degree) or the Department where a degree course is affiliated to. Quantitative measures of the clustering goodness are reported in Table <ref type="table">2</ref>, where the Silhouette Coefficients <ref type="bibr" target="#b31">[32]</ref> have been computed for each model, and again E5-mistral is the best performing one. In Appendix C, we report the scatter plots of the embedding spaces for each Embeddings LLM (Figure <ref type="figure" target="#fig_2">3</ref> and Figure <ref type="figure" target="#fig_3">4</ref> ). Plots have been obtained through a 2D dimensionality reduction using t-SNE <ref type="bibr" target="#b32">[33]</ref>.</p><p>We used the six QA test pairs to obtain also a quantitative evaluation of the correctness of the answers provided by all the Generation LLMs under investigation. Comparison was carried out against both the golden answers and the ones generated via gpt-3.5-turbo (GPT) in the original Unipa-GPT set up. The proposed evaluation task, can be regarded as an open QA one where, despite a golden answer is provided for a given question, diverse correct answers can be proposed with different linguistic nuances, according to Italian diaphasic variation <ref type="bibr" target="#b33">[34]</ref>. To evaluate both strict and light correctness of the generated answers, we employed traditional QA metrics such as BLEU <ref type="bibr" target="#b34">[35]</ref> (Figure <ref type="figure" target="#fig_1">2</ref>.a) and ROGUE-L score <ref type="bibr" target="#b35">[36]</ref> (Figure <ref type="figure" target="#fig_1">2</ref>.b) and novel metrics leveraging the RAGAS framework <ref type="bibr" target="#b30">[31]</ref> to evaluate Faithfulness (Figure <ref type="figure" target="#fig_1">2</ref>.c) and Correctness (Figure <ref type="figure" target="#fig_1">2</ref>.d) of the generated output. Such measures request an external LLM acting as a "judge", and we we used gpt-4-turbo in this respect. More specifically, Faithfulness measures the factual consistency of the generated answer against the given context, while Correctness involves gauging the accuracy of the generated answer when compared to the ground truth. Both metrics range from 0 to 1 and better performances are associated with higher scores.</p><p>Both BLEU and ROUGE scores are generally low, but we assume that this is mainly related to the fact that an exact match cannot be reached between the golden answer and the generated one, and a more semantically comparison should be taken into account. Overall, answers generated by gpt-3.5-turbo can be considered as the best ones as they attain highest values. By contrast, fine-tuning did not provided a desired improvement in the open-source models: all BLEU scores are almost zero, except for Anita-8B . ROUGE scores are higher than the corresponding BLEU ones, and again the base ver-</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Silhouette Coefficients for each Embedding LLM with reference to the two proposed clustering schemes, that is the degree courses typology and their affiliation to a particular Department. sion of each LLM performes better than the fine-tuned ones. Generally speaking, Anita-8B and Llama-3-8Binstruct outperform Minerva, since both reach comparable scores, but we assume that the tailored Italian fine-tuning over Llama-3 to obtain Anita-8B was crucial to make it the best performing open-source model during this first automatic evaluation phase. gpt-3.5-turbo exhibits the best Faithfulness scores despite being surpassed by Anita-8B in question Q2, and also these results confirm the previous considerations about BLEU scores. Something changes in evaluating models in terms of their Correctness: in this case gpt-3.5-turbo is the best model in three answers out of six, followed by Anita-8B (two best results) and Minerva-3B-20 (one best result). We are aware that gpt-based evaluation may lead to a preference over GPT models themselves, but gpt-4-turbo was the only high quality generative model we had access to at the time of making the experiments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Retriever</head><p>Overall results confirm that a (moderate) fine-tuning is not significantly beneficial in terms of performance increase for any model and, even if it does not reach the same performances, Anita-8B seems to be the most valuable alternative to GPT. A manual inspection of the generated answers, outlines a common issue related to the tokenization of the generated output: despite of its semantic correctness, the generated text is outputted as a unique word without any spaces, as Glielezionidelcorsosaracondottoattraversounaprocesso in Llama-3-8B-instruct, or it is over-splitted as <ref type="figure" target="#fig_2">e-domandre d'i-s-c-r-i-z-ion-e-per-l-A.-A.-2023-/--cor-so-n-d'-l-a-u-re-'-(M-agg</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>-is-t-ra-le)-a-dd-ac-ce-o-lib-ro</head><p>in Anita-8B. These errors make the models not suitable for human interaction, since it is not possible read the generated answers. We argue that a deeper analysis on the tokenizer that has been used and, a hyper-parametrs tuning in the generator, may lead to an increase of performances. Models tend also to answer in other languages as * La durada édié depresso àdue años, * Accesso libre! * Dipartment of Physics &amp; Chemistry "Emilo Segré" Codice course : 21915 in Llama-3-8B-instruct. We argue that this trouble can be related to the memory of multi-lingual models that uses texts also in French and Spanish despite the Italian fine-tuning. It is worth noticing that those languages are linguistically close to Italian and together belong to the Romance Languages <ref type="bibr" target="#b36">[37]</ref>. Thus, even if the output has to be considered wrong, a linguistic connection can be highlighted.</p><p>The most unsatisfactory results are reported for Minerva-3B: the model does not generate any answer related to the given question, and it seems that answers where generated with samples from model's training set. As stated before, a tuning of the generator hyperparameters may help in this case.</p><p>Despite the promising results, in some cases answers by both Anita-8B and Llama-3-8B-instruct are not good from a grammatical point of view, since they are full of mistakes, thus making them not yet ready to be used in real-world applications compared to OpenAI's ones.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions and future works</head><p>In this paper we presented Open Unipa-GPT, a virtual assistant, which is based solely on open-source LLMs, and uses a RAG approach to answer Italian universityrelated questions from secondary school students. The main intent of the presented research was setting up a sort of framework to test open-source small size LLMs, with either moderate or no fine-tuning at all, to be used for generating the embeddings and/or as text generation front-end in a RAG set up.</p><p>Our study led us to devise E5-mistral-7b-instruct as a valuable open-source alternative to OpenAI's embeddings, while none of the considered models attain a generation performance comparable to gpt-3.5-turbo, even after a fine-tuning procedure. The most promising Generation LLM, when plunged in our architecture, appears to be Anita-8B, but it still shows some issues related to both the tokenization and the grammatical correctness of the output. We are currently working to deep exploration of different fine-tuning approaches along with the use of huge size open-source LLMs for text generation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. unipa-corpus details</head><p>unipa-corpus <ref type="bibr" target="#b9">[10]</ref> is a collection of Italian documents that were retrieved directly from the website of the University of Palermo in Semptember 2023. The corpus is divided in two main sections, namely Education, that groups the available bachelor and master degree courses, and Future Students where important information about taxes payment and enrollment procedure are reported. For fine-tuning purposes, a semi-automatic procedure, involving gpt-3.5-turbo <ref type="bibr" target="#b2">[3]</ref>, was implemented to build a QA dataset. In Table <ref type="table" target="#tab_2">3</ref> are reported the statistics of unipa-corpus.    </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the Open Unipa-GPT architecture</figDesc><graphic coords="2,89.29,84.19,416.68,159.76" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Inference results over the generated answers according the following scores: (a) BLEU, (b) ROUGE, (c) Faithfulness and (d) Correctness. Due to displaying reasons, (a,b) are represented in a [0, 0.6] range, while (c,d) in a [0,1] range.</figDesc><graphic coords="5,89.29,202.10,416.69,209.79" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>C. Embedding spacesFigure 3 :</head><label>3</label><figDesc>Figure 3: Scatter plots of embedding spaces labeled as for typology</figDesc><graphic coords="11,89.29,117.83,416.68,202.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Scatter plots of embedding spaces labeled as for department</figDesc><graphic coords="11,89.29,365.71,416.68,207.43" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Context Relevancy scores over different Embedding LLMs. Bold values refer to the most relevant documents selected by RAGAS among the first four documents retrieved using the RAG. Underlined values refer to the golden documents.</figDesc><table><row><cell></cell><cell></cell><cell>Q1</cell><cell></cell><cell></cell><cell></cell><cell>Q2</cell><cell></cell><cell></cell><cell></cell><cell>Q3</cell><cell></cell><cell></cell></row><row><cell>model</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell></row><row><cell>open-ai-ada</cell><cell>0,1</cell><cell>0,1</cell><cell>0,0833</cell><cell>0,0714</cell><cell>0,0909</cell><cell>0,125</cell><cell>0,625</cell><cell>0,333</cell><cell>0,1</cell><cell>0,111</cell><cell>0,111</cell><cell>0,111</cell></row><row><cell>e5-mistral</cell><cell cols="4">0,0217 0,0345 0,0345 0,0233</cell><cell>0,5</cell><cell cols="2">0,0526 0,0333</cell><cell>0,025</cell><cell>0,0345</cell><cell>0,0154</cell><cell cols="2">0,0185 0,0435</cell></row><row><cell>bge</cell><cell>0,0345</cell><cell>0,0217</cell><cell cols="2">0,0385 0,0233</cell><cell>0,0526</cell><cell>0,312</cell><cell cols="2">0,0333 0,0909</cell><cell cols="3">0,0345 0,0667 0,0435</cell><cell>0,0154</cell></row><row><cell>bert-it</cell><cell>0,125</cell><cell>0,125</cell><cell>0,0345</cell><cell>0,0345</cell><cell>0,25</cell><cell>0,333</cell><cell>0,125</cell><cell>0,143</cell><cell>0,172</cell><cell>0,125</cell><cell>0,143</cell><cell>0,0192</cell></row><row><cell>m-e5</cell><cell>0,125</cell><cell>0,0217</cell><cell>0,1</cell><cell>0,0833</cell><cell>0,25</cell><cell>0,333</cell><cell>0,5</cell><cell>0,5</cell><cell>0,333</cell><cell>0,0185</cell><cell>0,037</cell><cell>0,0345</cell></row><row><cell></cell><cell></cell><cell>Q4</cell><cell></cell><cell></cell><cell></cell><cell>Q5</cell><cell></cell><cell></cell><cell></cell><cell>Q6</cell><cell></cell><cell></cell></row><row><cell>model</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell><cell>D1</cell><cell>D2</cell><cell>D3</cell><cell>D4</cell></row><row><cell>open-ai-ada</cell><cell>0,167</cell><cell>0,0588</cell><cell>0,1</cell><cell>0,333</cell><cell>0,429</cell><cell>0,5</cell><cell>0,143</cell><cell>0,111</cell><cell>1</cell><cell>0,1</cell><cell>0,167</cell><cell>0,5</cell></row><row><cell>e5-mistral</cell><cell>0,0417</cell><cell>0,05</cell><cell>0,05</cell><cell>0,276</cell><cell>0,154</cell><cell>0,04</cell><cell>0,5</cell><cell>0,0667</cell><cell>0,333</cell><cell>0,111</cell><cell>0,333</cell><cell>0,111</cell></row><row><cell>bge</cell><cell>0,241</cell><cell>0,0417</cell><cell>0,333</cell><cell>0,1</cell><cell>0,154</cell><cell>0,04</cell><cell>0,333</cell><cell>0,05</cell><cell>0,182</cell><cell>0,111</cell><cell>0,444</cell><cell>0,333</cell></row><row><cell>bert-it</cell><cell>0,152</cell><cell>0,0303</cell><cell>0,152</cell><cell>0,0303</cell><cell>0,5</cell><cell>0,04</cell><cell cols="2">0,0385 0,0769</cell><cell>0,111</cell><cell>0,333</cell><cell>0,25</cell><cell>0,25</cell></row><row><cell>m-e5</cell><cell>0,333</cell><cell>0,5</cell><cell>0,5</cell><cell>0,5</cell><cell>0,154</cell><cell>0,333</cell><cell>0,167</cell><cell>0,5</cell><cell>0,333</cell><cell>0,5</cell><cell>0,111</cell><cell>0,333</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Number of documents and QA pairs in unipa-corpus.</figDesc><table><row><cell></cell><cell>Education</cell><cell>Future Students</cell></row><row><cell>Documents</cell><cell>506</cell><cell>104</cell></row><row><cell>Tokens</cell><cell>1072214</cell><cell>987424</cell></row><row><cell>QA pairs train</cell><cell>506</cell><cell>269</cell></row><row><cell>Tokens train</cell><cell>191612</cell><cell>68160</cell></row><row><cell>QA pairs val</cell><cell>253</cell><cell>133</cell></row><row><cell>Tokens val</cell><cell>93443</cell><cell>29675</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Overview of 6 QA pairs manually generated used for evaluation purposes Il corso di laurea magistrale in Chimica presso l'Università degli Studi di Palermo, ha sede a Palermo e ha una durata di 2 anni. L'obiettivo del corso è la formazione di laureati con una solida preparazione di base, che li ponga in grado di affrontare con competenza ed autonomia sia attività professionali sia quelle di ricerca accademica ed industriale. Per questo motivo il corso di studio è progettato con lo scopo di completare la cultura scientifica nelle discipline chimiche fondamentali e di introdurre conoscenze più avanzate. Quindi lo studente può completare il proprio piano di studi, attenzionando gli insegnamenti che meglio si adattano ai propri interessi specifici, come la chimica dei materiali, dell'ambiente e dei beni culturali, la chimica supramolecolare e la didattica della chimica. Per maggiori informazioni vai su http://www.unipa.it/struttura.html?id=766</figDesc><table><row><cell>IDs</cell><cell>Questions</cell><cell>Answers</cell></row><row><cell>Q1</cell><cell>Chi è il professore di Intelligenza Artificiale 1 per il corso di Laurea Magistrale in Ingegneria Informatica?</cell><cell>Il professore di Intelligenza Artificiale 1 del corso di Laurea Magistrale in Ingegneria Informatica è il professore Gaglio e l'insegnamento verrà erogato durante il primo semestre. Per maggiori informazioni vai su http://www.unipa.it/struttura.html?id=721</cell></row><row><cell></cell><cell></cell><cell>La presentazione delle domande di iscrizione per l'Anno Accademico</cell></row><row><cell></cell><cell></cell><cell>2023/2024 varia in base alla tipologia di corso. Per i corsi di studio</cell></row><row><cell></cell><cell></cell><cell>triennali e magistrali a ciclo unico, la domanda può essere presentata</cell></row><row><cell>Q2</cell><cell>Quali sono le scadenze di iscrizione ad un corso di laurea?</cell><cell>dall'1 agosto al 30 settembre 2023 , mentre per i corsi magistrali, dall'1 agosto al 30 novembre 2023, in entrambi i casi, è richiesto il pagamento della prima rata delle tasse universitarie. Per i corsi di studio ad</cell></row><row><cell></cell><cell></cell><cell>accesso programmato locale, con prenotazione online o ad accesso</cell></row><row><cell></cell><cell></cell><cell>programmato nazionale, le immatricolazioni vanno effettuate entro i</cell></row><row><cell></cell><cell></cell><cell>termini previsti dal relativi bandi di concorso.</cell></row><row><cell>Q3</cell><cell>Come funziona la magistrale di chimica?</cell><cell></cell></row><row><cell></cell><cell></cell><cell>Ciao! Il corso di Laurea in Consulente Giuridico d'Impresa sembra</cell></row><row><cell>Q4</cell><cell>ciao! sono un ragazzo appena uscito dal liceo che è interessato al settore legale,in particolare alle leggi sulle aziende. dove potrei iscrivermi?</cell><cell>proprio fare al caso tuo! Il corso di laurea ha sede a Trapani e ha una durata di 3 anni e ti fornirà le conoscenze in ambito giuridico ed economico-aziendalistico. Una volta terminato il percorso di studio potrai svolgere attività interdisciplinari che richiedono competenze giuridiche, aziendalistiche e organizzativo-gestionali. Per maggiori</cell></row><row><cell></cell><cell></cell><cell>informazioni vai su http://www.unipa.it/struttura.htmlid=1557</cell></row><row><cell></cell><cell></cell><cell>È possibile recarsi in segreteria il lunedì, mercoledì e venerdì dalle 10.00</cell></row><row><cell>Q5</cell><cell>come posso prenotare un appuntamento in segreteria?</cell><cell>alle 12.00, martedì e giovedì dalle 15.00 alle 17.00 . Puoi prenotare il tuo turno attraverso la App SolariQ. Per maggiori informazioni vai su</cell></row><row><cell></cell><cell></cell><cell>https://www.unipa.it/servizi/segreterie/</cell></row><row><cell></cell><cell></cell><cell>Il pagamento delle tasse deve essere effettuato esclusivamente mediante</cell></row><row><cell></cell><cell></cell><cell>sistema PAgoPA (Pagamenti della Pubblica Amministrazione). Dopo aver</cell></row><row><cell></cell><cell></cell><cell>compilato la pratica online, è possibile pagare direttamente online con</cell></row><row><cell></cell><cell></cell><cell>il sistema PAgoPA o stampare il bollettino e pagare presso tabaccai</cell></row><row><cell>Q6</cell><cell>Come si pagano le tasse?</cell><cell>convenzionati o ricevitorie abilitate PAgoPA. Ulteriori informazioni sul pagamento via PAgoPA sono reperibili qui https://immaweb.unipa.it/</cell></row><row><cell></cell><cell></cell><cell>immaweb/public/pagamenti.seam, mentre è disponibile il Regolamento in</cell></row><row><cell></cell><cell></cell><cell>materia di contribuzione studentesca https://www.unipa.it/servizi/segreterie/</cell></row><row><cell></cell><cell></cell><cell>.content/documenti/regolamenti_calendari/2023/5105144-</cell></row><row><cell></cell><cell></cell><cell>def_regolamento-contribuzione-studentesca-2023-24-2.pdf</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>English version of Table4. The Master's degree course in Chemistry at the University of Palermo is based in Palermo and lasts 2 years. The aim of the course is to train graduates with a good background, enabling them to deal competently and independently with both professional activities and academic and industrial research. For this reason, the course is designed with the aim of completing the scientific culture in the fundamental chemical disciplines and introducing more advanced knowledge. Therefore, students can complete their study plan by focusing on the subjects that best suit their specific interests, such as the chemistry of materials, the environment and cultural heritage, supramolecular chemistry and the didactics of chemistry. For more information go to http://www.unipa.it/struttura.html?id=766 Hi! The Bachelor of Business Law Consultant programme sounds like it could be just the thing for you! The degree course is based in Trapani and lasts 3 years and will provide you with knowledge in the fields of law and business economics. Once you have completed the course you will be able to carry out interdisciplinary activities requiring legal, business and organisational-managerial skills. For more information go to http://www.unipa.it/struttura.html?id=1557You can go to the secretariat on Mondays, Wednesdays and Fridays from 10 a.m. to 12 noon, Tuesdays and Thursdays from 3 p.m. to 5 p.m. . You can book your appointment through the SolariQ App. For more information go to https://www.unipa.it/servizi/segreterie/ Fees must be paid exclusively through the PAgoPA (Public Administration Payments) system, which is accessed through the university portal. After completing the paperwork online, you can either pay directly online via the PAgoPA system or print out the payment slip and pay at a PAgoPA-enabled tax office. Further information on paying via PAgoPA can be found here https://immaweb.unipa.it/immaweb/public/pagamenti. seam, while the Student Contribution Regulations is available here https://www.unipa.it/servizi/segreterie/.content/documents/regulations_ calendars/2023/5105144-def_regulation-student-contribution-2023-24-2.pdf</figDesc><table><row><cell>IDs</cell><cell>Questions</cell><cell>Answers</cell></row><row><cell></cell><cell>Who is the Artificial</cell><cell>The Artificial Intelligence 1 professor for the Computer Engineering</cell></row><row><cell>Q1</cell><cell>Intelligence 1 professor for Computer Engineering</cell><cell>Master degree course is Professor Gaglio and it will be delivered during the first semester. For more information go to</cell></row><row><cell></cell><cell>Master degree course?</cell><cell>http://www.unipa.it/struttura.html?id=721</cell></row><row><cell></cell><cell></cell><cell>The submission of applications for the Academic Year 2023/2024 varies</cell></row><row><cell></cell><cell></cell><cell>according to the type of course. For three-year and single-cycle</cell></row><row><cell></cell><cell></cell><cell>master's degree courses, applications can be submitted from 1 August</cell></row><row><cell></cell><cell>What are the deadlines</cell><cell>to 30 September 2023, while for master's degree courses, from 1 August</cell></row><row><cell>Q2</cell><cell>for enrolling in a</cell><cell>to 30 November 2023; in both payment of the first instalment of</cell></row><row><cell></cell><cell>degree programme?</cell><cell>tuition fees is For courses with local programmed access,</cell></row><row><cell></cell><cell></cell><cell>with online booking or national programmed access, enrolment must be</cell></row><row><cell></cell><cell></cell><cell>carried out by the deadlines set out in the corresponding calls for</cell></row><row><cell></cell><cell></cell><cell>application.</cell></row><row><cell>Q3</cell><cell>How does the master's degree in chemistry work?</cell><cell></cell></row><row><cell></cell><cell>hello! I'm a guy just out of high</cell><cell></cell></row><row><cell>Q4</cell><cell>school who is interested in law, especially corporate law.</cell><cell></cell></row><row><cell></cell><cell>where should i apply?</cell><cell></cell></row><row><cell>Q5</cell><cell>how can i book an appointment at the secretariat?</cell><cell></cell></row><row><cell>Q6</cell><cell>How do I pay fees?</cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://www.langchain.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">as in https://huggingface.co/spaces/mteb/leader-board in June 2024</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://huggingface.co/nickprock/ sentence-bert-base-italian-xxl-uncased</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">https://github.com/tloen/alpaca-lora</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Stone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Albert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Almahairi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Babaei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bashlykov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Batra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bhargava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bhosale</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2307.09288</idno>
		<title level="m">Llama 2: Open foundation and fine-tuned chat models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">The llama 3 herd of models</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">. M</forename><surname>Llama</surname></persName>
		</author>
		<author>
			<persName><surname>Team</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2407.21783.arXiv:2407.21783" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Language models are few-shot learners</title>
		<author>
			<persName><forename type="first">T</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Kaplan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="1877" to="1901" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2303.08774.arXiv:2303.08774" />
		<title level="m">Gpt-4</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">technical report</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-Z</forename><forename type="middle">J</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">L</forename><surname>Wang</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2312.12148.arXiv:2312.12148" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wallis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Allen-Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2106.09685</idno>
		<title level="m">Lora: Low-rank adaptation of large language models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">-bit matrix multiplication for transformers at scale</title>
		<author>
			<persName><forename type="first">T</forename><surname>Dettmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Belkada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2208.07339.arXiv:2208.07339" />
	</analytic>
	<monogr>
		<title level="m">Llm.int8</title>
				<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page">8</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2405.07101</idno>
		<title level="m">Advanced natural-based interaction for the italian language: Llamantino-3-anita</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Orlando</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Moroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-L</forename><surname>Huguet Cabot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Conia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Barba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Navigli</surname></persName>
		</author>
		<ptr target="https://nlp.uniroma1.it/minerva/" />
		<title level="m">Minerva technical report</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Siragusa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pirrone</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2407.14246.arXiv:2407.14246" />
		<title level="m">Unipa-gpt: Large language models for university-oriented qa in italian</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Retrieval-augmented generation for knowledge-intensive nlp tasks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Karpukhin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Küttler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>-T. Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rocktäschel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="9459" to="9474" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1423</idno>
		<ptr target="https://aclanthology.org/N19-1423.doi:10.18653/v1/N19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Burstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Doran</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Solorio</surname></persName>
		</editor>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Alberto: Italian bert language understanding model for nlp challenging tasks based on tweets</title>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Degemmis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:204914950" />
	</analytic>
	<monogr>
		<title level="m">Italian Conference on Computational Linguistics</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Musacchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siciliani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fiameni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.09993</idno>
		<title level="m">Llamantino: Llama 2 models for effective text generation in italian language</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Bacciu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Trappolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Santilli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rodolà</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Silvestri</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2306.14457</idno>
		<title level="m">Fauno: The italian large language model that will leave you senza parole!</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Santilli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rodolà</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2307.16456</idno>
		<title level="m">Camoscio: an italian instruction-tuned llama</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Ai4la: An intelligent chatbot for supporting students with dyslexia, based on generative ai</title>
		<author>
			<persName><forename type="first">S</forename><surname>D'urso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Sciarrone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Generative Intelligence and Intelligent Tutoring Systems</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Sifaleras</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Lin</surname></persName>
		</editor>
		<meeting><address><addrLine>Nature Switzerland, Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="369" to="377" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Retrieval augmented generation (rag) based restaurant chatbot with ai testability</title>
		<author>
			<persName><forename type="first">V</forename><surname>Bhat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sree</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cheerla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mathew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Reinforcement learning for optimizing rag for domain chatbots</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kulkarni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tangarajan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Trivedi</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2401.06800.arXiv:2401.06800" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Two-phase ragbased chatbot for italian funding application assistance</title>
		<author>
			<persName><forename type="first">T</forename><surname>Boccato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ferrante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Toschi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A retrieval-augmented generation strategy to enhance medical chatbot reliability</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Ghanbari</forename><surname>Haez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Segala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bellan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Magnolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sanna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Consolandi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dragoni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Artificial Intelligence in Medicine</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Finkelstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Moskovitch</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Parimbelli</surname></persName>
		</editor>
		<meeting><address><addrLine>Nature Switzerland, Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="213" to="223" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Towards an llm-based intelligent assistant for industry 5.0</title>
		<author>
			<persName><forename type="first">R</forename><surname>Figliè</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Turchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Baldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mazzei</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3701/paper7.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st International Workshop on Designing and Building Hybrid Human-AI Systems</title>
				<meeting>the 1st International Workshop on Designing and Building Hybrid Human-AI Systems<address><addrLine>SYNERGY</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
			<biblScope unit="volume">3701</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Billion-scale similarity search with GPUs</title>
		<author>
			<persName><forename type="first">J</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Douze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jégou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Big Data</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="535" to="547" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Muennighoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tazi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Magne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2210.07316</idno>
		<idno type="arXiv">arXiv:2210.07316</idno>
		<ptr target="https://arxiv.org/abs/2210.07316.doi:10.48550/ARXIV.2210.07316" />
		<title level="m">Mteb: Massive text embedding benchmark</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2402.03216.arXiv:2402.03216" />
		<title level="m">Bge m3-embedding: Multi-lingual, multi-functionality, multi-granularity text embeddings through self-knowledge distillation</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Majumder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wei</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.00368</idno>
		<title level="m">Improving text embeddings with large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Majumder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wei</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2402.05672</idno>
		<title level="m">Multilingual e5 text embeddings: A technical report</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ł</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title/>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Q</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sablayrolles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mensch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bamford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Chaplot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>De Las Casas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bressand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lengyel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Saulnier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Lavaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Sayed</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2310.06825.arXiv:2310.06825" />
	</analytic>
	<monogr>
		<title level="j">Mistral</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Efficiency-driven custom chatbot development: Unleashing langchain, rag, and performanceoptimized llm fusion</title>
		<author>
			<persName><forename type="first">S</forename><surname>Vidivelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ramachandran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dharunbalaji</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers, Materials &amp; Continua</title>
		<imprint>
			<biblScope unit="volume">80</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Es</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>James</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa-Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schockaert</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.15217</idno>
		<title level="m">Ragas: Automated evaluation of retrieval augmented generation</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Silhouettes: a graphical aid to the interpretation and validation of cluster analysis</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Rousseeuw</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of computational and applied mathematics</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="53" to="65" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Visualizing data using t-sne</title>
		<author>
			<persName><forename type="first">L</forename><surname>Van Der Maaten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of machine learning research</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Berruto</surname></persName>
		</author>
		<ptr target="https://www.treccani.it/enciclopedia/variazione-diafasica_(Enciclopedia-dell&apos;Italiano)/" />
		<title level="m">Variazione diafasica</title>
				<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Bleu: a method for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Papineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roukos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-J</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th annual meeting of the Association for Computational Linguistics</title>
				<meeting>the 40th annual meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="311" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Rouge: A package for automatic evaluation of summaries</title>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Text summarization branches out</title>
				<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="74" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Alkire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Rosen</surname></persName>
		</author>
		<title level="m">Romance languages: A historical introduction</title>
				<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
