<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Unipa-GPT: a framework to assess open-source alternatives to Chat-GPT for Italian chat-bots</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Irene Siragusa</string-name>
          <email>irene.siragusa02@unipa.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberto Pirrone</string-name>
          <email>roberto.pirrone@unipa.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, IT University of Copenhagen</institution>
          ,
          <addr-line>København S, 2300</addr-line>
          ,
          <country country="DK">Denmark</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Engineering, University of Palermo</institution>
          ,
          <addr-line>Palermo, 90128, Sicily</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>RAG</institution>
          ,
          <addr-line>ChatGPT, LLM, Embedding</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>The increasing interest in developing Language Models</institution>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Unipa-GPT is described in Section 3</institution>
          ,
          <addr-line>and an overview</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper illustrates the implementation of Open Unipa-GPT, an open-source version of the Unipa-GPT chat-bot that leverages open-source Large Language Models for embeddings and text generation. The system relies on a Retrieval Augmented Generation approach, thus mitigating hallucination errors in the generation phase. A detailed comparison between diferent models is reported to illustrate their performance as regards embedding generation, retrieval, and text generation. In the last case, models were tested in a simple inference setup after a fine-tuning procedure. Experiments demonstrate that an open-source LLMs can be eficiently used for embedding generation, but none of the models does reach the performances obtained by closed models, such as gpt-3.5-turbo in generating answers. Corpora and code are available on GitHub1 Workshop Proceedings available also in the 8B and 13B parameters versions, in our original version, namely text-embedding-adacial for developing Natural Language Process (NLP) ap- and related results are reported in Section 5. Finally,</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR</p>
      <p>
        ceur-ws.org
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Anita [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] are a fine-tuned Italian version of the instruct
Attribution 4.0 International (CC BY 4.0).
      </p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <sec id="sec-2-1">
        <title>The increasing development of bigger and bigger Large</title>
      </sec>
      <sec id="sec-2-2">
        <title>Language Models (LLM), reaching 70B parameters as for</title>
      </sec>
      <sec id="sec-2-3">
        <title>Meta LLMs (Llama 2 [1] and Llama 3 [2]) and more as for</title>
        <p>
          OpenAI ones (GPT-3 [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] and GPT-4 [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]1), requires a
significant computational resources for training, fine-tuning
or inference. OpenAI models are accessible only upon
payment via OpenAI API and cannot be downloaded
in any way, while the open-source models by Meta are
and they can either be fine-tuned via Parameter-Eficient
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Fine-Tuning techniques (PEFT) [5] such as LoRA [6], or</title>
        <p>
          they can make direct inference using a 8-bit quantization
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] keeping the computational resources relatively small.
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>The availability of open-source small-size LLMs is cru</title>
        <p>
          plications that leverage a fine-tuning procedure over a
specific domain or language, as for Anita [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], an Italian
        </p>
      </sec>
      <sec id="sec-2-6">
        <title>8B adaptation of Llama 3.</title>
        <p>
          Nevertheless, GPT and Llama models cannot be
considered as truly open-source since their training data set
is not available and, as for GPT models, and also their
actual architecture is not accessible. Minerva [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] model, on
tecture, weights, and training data are accessible, but it
nEvelop-O
(R. Pirrone)
Dec 04 — 06, 2024, Pisa, Italy
∗Corresponding author.
(LM) for the Italian language, starts when BERT [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] was
were developed. After ChatGPT was made public [
          <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
          ],
an increasing interest in developing and using LLMs, and
cial, also for the Italian NLP community, thus leading to
the development of foundational models based on Llama
        </p>
      </sec>
      <sec id="sec-2-7">
        <title>2 [1] and Llama 3 [2]. Among those models, LLaMantino</title>
        <p>(chat version) [14] and Fauno [15], are based on Llama</p>
      </sec>
      <sec id="sec-2-8">
        <title>2 fine-tuned for chat purposes, while Camoscio [16] and</title>
        <p>version of Llama 2 and Llama 3, respectively.
the other side, is an Italian and English LLM whose archi- first released and adapted models, such as AlBERTo [ 13]
CLiC-it 2024: Tenth Italian Conference on Computational Linguistics, in generative AI based on decoder-only model, was
cru</p>
        <p>RAG is used in developing chat-bots which are 1K token chunks with an overlap of 50 tokens. Split
docgrounded in various domains where the models need uments are then processed by a LLM (the Embedding
to be deeply guided in generation to avoid hallucination LLM ) to generate the corresponding embedding, and
in their answers. Various examples can be found in the store them in the vector database. Diferent LLMs were
educational domain as for AI4LA [17], an assistant to used for embedding generation: we selected the best
modstudents with Specific Learning Disorders (SLDs) like els according to the Massive Text Embedding Benchmark
Dyslexia, Dysorthographia, and Dyscalculia, or as as- (MTEB) [24] for Information Retrieval3. We selected only
sistant providing information about restaurant industry models that explicitly state that they were trained and
[18] or as chat-bot for Frequently Asked Questions (FAQ) tested also with Italian data. In the end, we selected
[19]. Also chat-bots for the Italian language were imple- the following models: BGE-M3 (BGE) [25],
E5-mistralmented for real-wold applications, namely as assistant 7b-instruct (E5-mistral) [26],
sentence-bert-basefor Italian Funding Application [20], or in the medical italian-xxl-uncased4 (BERT-it) and
Multilingualdomain [21] or in industrial context [22]. The aforemen- E5-large-instruct (m-E5) [27] .
tioned works share the same architecture with the one A vector database was built for each model, and their
we used to implement our model. In contrast with them, corresponding embedding spaces were compared to each
we decided to stress capabilities of open-source LLMs other and with text-embedding-ada-002, the
embedand do not rely on GPT-based models, that are used as ding model from OpenAI, to asses their retrieval
perforbaseline reference for text generation (gpt-3.5-turbo) mances (Section 5).
and as an external judge to evaluate performances of the
other models (gpt-4.5-turbo).</p>
        <sec id="sec-2-8-1">
          <title>3.2. Generator</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. System architecture</title>
      <sec id="sec-3-1">
        <title>Open Unipa-GPT relies on two main components as it is</title>
        <p>shown in Figure 1 that is the Retriever and the Generator.
In the following, the two components are detailed.</p>
        <sec id="sec-3-1-1">
          <title>3.1. Retriever</title>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>The Retriever is made up of a vector database built us</title>
        <p>ing the LangChain framework2, which makes use of the
Facebook AI Similarity Search (FAISS) library [23]. The
vector database is filled with the documents belonging
to the unipa-corpus (Appendix A), that are divided into</p>
      </sec>
      <sec id="sec-3-3">
        <title>2https://www.langchain.com</title>
      </sec>
      <sec id="sec-3-4">
        <title>The Generator uses the following Italian isntruction prompt to answer to user questions:</title>
        <p>
          Sei Unipa-GPT, chatbot e assistente virtuale
dell’Università degli Studi di Palermo che
risponde cordialmente e in forma
colloquiale. Ai saluti, rispondi salutando e
presentandoti. Ricordati che il rettore
dell’Università è il professore Massimo Midiri. Se la
domanda riguarda l’università degli studi di
Palermo, rispondi in base alle informazioni
e riporta i link ad esse associati; Se non
3as in https://huggingface.co/spaces/mteb/leader-board in June 2024
4https://huggingface.co/nickprock/
sentence-bert-base-italian-xxl-uncased
The proposed model is intended to work in an open QA
context, where correct answers are not known, thus, after
a previous phase of qualitative evaluation [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] as in [17,
20, 21, 22], we opted for a quantitative analysis, relying
on the small QA data set described in Section 4 to evaluate
the performances against a set of golden labels in terms
of both retrieval and answering capabilities [30, 19, 18].
        </p>
        <p>For each QA test pair, we retrieved the four most
relevant documents from each vector database related
to one of the open Embedding LLMs under
investiga</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. The data set</title>
      <sec id="sec-4-1">
        <title>The Italian documents data set built for Unipa-GPT is</title>
        <p>
          called unipa-corpus [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], and it has been generated
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>5https://github.com/tloen/alpaca-lora</title>
        <p>tion. Then we scored the retrieved documents in terms dimensionality reduction using t-SNE [33].
of their context relevancy with respect to the provided We used the six QA test pairs to obtain also a
quantitaquestion using the RAGAS framework [31] that exploits tive evaluation of the correctness of the answers provided
gpt-4-turbo for the evaluation. Results are reported by all the Generation LLMs under investigation.
Comin Table 1, and they include also the performances of parison was carried out against both the golden answers
the original vector database using OpenAI embeddings and the ones generated via gpt-3.5-turbo (GPT) in the
(text-embedding-ada-002, referred as open-ai-ada). original Unipa-GPT set up. The proposed evaluation
The overall scores are not so high, and also the highest task, can be regarded as an open QA one where, despite
relevancy do not always correspond to the golden docu- a golden answer is provided for a given question, diverse
ment used for generating the corresponding answer. In correct answers can be proposed with diferent linguistic
Table 1 the underlined values are the ones associated with nuances, according to Italian diaphasic variation [34]. To
golden documents, while the bold ones are the highest evaluate both strict and light correctness of the generated
RAGAS values. A model is considered to perform cor- answers, we employed traditional QA metrics such as
rectly if the highest context relevancy score is assigned BLEU [35] (Figure 2.a) and ROGUE-L score [36] (Figure
to one of the golden documents. This evaluation pro- 2.b) and novel metrics leveraging the RAGAS framework
cedure led to select E5-mistral as the best performing [31] to evaluate Faithfulness (Figure 2.c) and Correctness
Embeddings LLM among the ones we investigated. (Figure 2.d) of the generated output. Such measures
re</p>
        <p>Superior performances of E5-mistral are also con- quest an external LLM acting as a “judge”, and we we used
ifrmed by a deep analysis on the embeddings space by gpt-4-turbo in this respect. More specifically,
Faithfulmeans of two diferent clustering procedures. We clus- ness measures the factual consistency of the generated
tered the embeddings generated by each LLM starting answer against the given context, while Correctness
infrom the documents belonging to both sections Educa- volves gauging the accuracy of the generated answer
tional Ofer and Future Students of the UniPA website. when compared to the ground truth. Both metrics range
The firs group of documents is the list of all the available from 0 to 1 and better performances are associated with
courses at the University, while the second group con- higher scores.
tains useful information for future students who want to Both BLEU and ROUGE scores are generally low, but
enroll in a degree course. We clustered the embedding we assume that this is mainly related to the fact that
spaces according to the either the course degree typol- an exact match cannot be reached between the golden
ogy (bachelor/master degree) or the Department where answer and the generated one, and a more semantically
a degree course is afiliated to. Quantitative measures of comparison should be taken into account. Overall,
anthe clustering goodness are reported in Table 2, where swers generated by gpt-3.5-turbo can be considered as
the Silhouette Coeficients [ 32] have been computed for the best ones as they attain highest values. By contrast,
each model, and again E5-mistral is the best performing ifne-tuning did not provided a desired improvement in
one. In Appendix C, we report the scatter plots of the the open-source models: all BLEU scores are almost zero,
embedding spaces for each Embeddings LLM (Figure 3 except for Anita-8B . ROUGE scores are higher than
and Figure 4 ). Plots have been obtained through a 2D the corresponding BLEU ones, and again the base
version of each LLM performes better than the fine-tuned is not significantly beneficial in terms of performance
ones. Generally speaking, Anita-8B and Llama-3-8B- increase for any model and, even if it does not reach
instruct outperform Minerva, since both reach com- the same performances, Anita-8B seems to be the most
parable scores, but we assume that the tailored Italian valuable alternative to GPT.
ifne-tuning over Llama-3 to obtain Anita-8B was crucial A manual inspection of the generated answers,
outto make it the best performing open-source model during lines a common issue related to the tokenization of the
this first automatic evaluation phase. generated output: despite of its semantic correctness, the
gpt-3.5-turbo exhibits the best Faithfulness scores generated text is outputted as a unique word without any
despite being surpassed by Anita-8B in question Q2, and spaces, as
also these results confirm the previous considerations
about BLEU scores. Something changes in evaluating
Glielezionidelcorsosaracondottoattravermodels in terms of their Correctness: in this case gpt- sounaprocesso
3.5-turbo is the best model in three answers out of six, in Llama-3-8B-instruct, or it is over-splitted as
followed by Anita-8B (two best results) and
Minerva3B-20 (one best result). We are aware that gpt-based e-domandre
d’i-s-c-r-i-z-ion-e-per-l-A.evaluation may lead to a preference over GPT models
A.–2023–/—-cor-so-n-d’-l-a-u-re-’-(M-agthemselves, but gpt-4-turbo was the only high quality g-is-t-ra-le)–a-dd-ac-ce-o-lib-ro
generative model we had access to at the time of making
the experiments.</p>
        <p>Overall results confirm that a (moderate) fine-tuning
in Anita-8B. These errors make the models not suitable
for human interaction, since it is not possible read the
generated answers. We argue that a deeper analysis on
the tokenizer that has been used and, a hyper-parametrs
tuning in the generator, may lead to an increase of
performances. Models tend also to answer in other languages
as
* La durada édié depresso àdue años, *
Accesso libre! * Dipartment of Physics &amp;
Chemistry “Emilo Segré” Codice course :
21915
in Llama-3-8B-instruct. We argue that this trouble
can be related to the memory of multi-lingual models that
uses texts also in French and Spanish despite the Italian
ifne-tuning. It is worth noticing that those languages are
linguistically close to Italian and together belong to the
Romance Languages [37]. Thus, even if the output has
to be considered wrong, a linguistic connection can be
highlighted.</p>
        <p>The most unsatisfactory results are reported for
Minerva-3B: the model does not generate any answer
related to the given question, and it seems that answers
where generated with samples from model’s training
set. As stated before, a tuning of the generator
hyperparameters may help in this case.</p>
        <p>Despite the promising results, in some cases answers
by both Anita-8B and Llama-3-8B-instruct are not
good from a grammatical point of view, since they are
full of mistakes, thus making them not yet ready to be
used in real-world applications compared to OpenAI’s
ones.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Conclusions and future works</title>
      <p>In this paper we presented Open Unipa-GPT, a virtual
assistant, which is based solely on open-source LLMs,
and uses a RAG approach to answer Italian
universityrelated questions from secondary school students. The
main intent of the presented research was setting up a
sort of framework to test open-source small size LLMs,
with either moderate or no fine-tuning at all, to be used
for generating the embeddings and/or as text generation
front-end in a RAG set up.</p>
      <p>Our study led us to devise E5-mistral-7b-instruct
as a valuable open-source alternative to OpenAI’s
embeddings, while none of the considered models attain a
generation performance comparable to gpt-3.5-turbo,
even after a fine-tuning procedure. The most
promising Generation LLM, when plunged in our architecture,
appears to be Anita-8B, but it still shows some issues
related to both the tokenization and the grammatical
correctness of the output. We are currently working
to deep exploration of diferent fine-tuning approaches
along with the use of huge size open-source LLMs for
text generation.
https://aclanthology.org/N19-1423. doi:10.18653/ [25] J. Chen, S. Xiao, P. Zhang, K. Luo, D. Lian,
v1/N19- 1423. Z. Liu, Bge m3-embedding: Multi-lingual,
[13] M. Polignano, P. Basile, M. Degemmis, G. Semeraro, multi-functionality, multi-granularity text
emV. Basile, Alberto: Italian bert language under- beddings through self-knowledge distillation,
standing model for nlp challenging tasks based on 2024. URL: https://arxiv.org/abs/2402.03216.
tweets, in: Italian Conference on Computational arXiv:2402.03216.</p>
      <p>Linguistics, 2019. URL: https://api.semanticscholar. [26] L. Wang, N. Yang, X. Huang, L. Yang, R. Majumder,
org/CorpusID:204914950. F. Wei, Improving text embeddings with large
lan[14] P. Basile, E. Musacchio, M. Polignano, L. Siciliani, guage models, arXiv preprint arXiv:2401.00368
G. Fiameni, G. Semeraro, Llamantino: Llama 2 mod- (2023).
els for efective text generation in italian language, [27] L. Wang, N. Yang, X. Huang, L. Yang, R. Majumder,
2023. arXiv:2312.09993. F. Wei, Multilingual e5 text embeddings: A
techni[15] A. Bacciu, G. Trappolini, A. Santilli, E. Rodolà, F. Sil- cal report, arXiv preprint arXiv:2402.05672 (2024).
vestri, Fauno: The italian large language model [28] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit,
that will leave you senza parole!, arXiv preprint L. Jones, A. N. Gomez, Ł. Kaiser, I. Polosukhin,
AtarXiv:2306.14457 (2023). tention is all you need, Advances in neural
infor[16] A. Santilli, E. Rodolà, Camoscio: an italian mation processing systems 30 (2017).</p>
      <p>instruction-tuned llama, 2023. arXiv:2307.16456. [29] A. Q. Jiang, A. Sablayrolles, A. Mensch, C.
Bam[17] S. D’Urso, F. Sciarrone, Ai4la: An intelligent chat- ford, D. S. Chaplot, D. de las Casas, F. Bressand,
bot for supporting students with dyslexia, based on G. Lengyel, G. Lample, L. Saulnier, L. R. Lavaud,
M.generative ai, in: A. Sifaleras, F. Lin (Eds.), Gen- A. Lachaux, P. Stock, T. L. Scao, T. Lavril, T. Wang,
erative Intelligence and Intelligent Tutoring Sys- T. Lacroix, W. E. Sayed, Mistral 7b, 2023. URL: https:
tems, Springer Nature Switzerland, Cham, 2024, pp. //arxiv.org/abs/2310.06825. arXiv:2310.06825.
369–377. [30] S. Vidivelli, M. Ramachandran, A.
Dharunbal[18] V. Bhat, D. Sree, J. Cheerla, N. Mathew, G. LIu, J. Gao, aji, Eficiency-driven custom chatbot
developRetrieval augmented generation (rag) based restau- ment: Unleashing langchain, rag, and
performancerant chatbot with ai testability, 2024. optimized llm fusion., Computers, Materials &amp;
Con[19] M. Kulkarni, P. Tangarajan, K. Kim, A. Trivedi, Re- tinua 80 (2024).</p>
      <p>inforcement learning for optimizing rag for do- [31] S. Es, J. James, L. Espinosa-Anke, S. Schockaert,
Ramain chatbots, 2024. URL: https://arxiv.org/abs/ gas: Automated evaluation of retrieval augmented
2401.06800. arXiv:2401.06800. generation, arXiv preprint arXiv:2309.15217 (2023).
[20] T. Boccato, M. Ferrante, N. Toschi, Two-phase rag- [32] P. J. Rousseeuw, Silhouettes: a graphical aid to
based chatbot for italian funding application assis- the interpretation and validation of cluster analysis,
tance, 2024. Journal of computational and applied mathematics
[21] S. Ghanbari Haez, M. Segala, P. Bellan, S. Mag- 20 (1987) 53–65.</p>
      <p>nolini, L. Sanna, M. Consolandi, M. Dragoni, A [33] L. Van der Maaten, G. Hinton, Visualizing data
retrieval-augmented generation strategy to en- using t-sne., Journal of machine learning research
hance medical chatbot reliability, in: J. Finkelstein, 9 (2008).</p>
      <p>R. Moskovitch, E. Parimbelli (Eds.), Artificial Intel- [34] G. Berruto, Variazione diafasica, 2011.
ligence in Medicine, Springer Nature Switzerland, URL: https://www.treccani.it/enciclopedia/
Cham, 2024, pp. 213–223. variazione-diafasica_(Enciclopedia-dell'Italiano)/.
[22] R. Figliè, T. Turchi, G. Baldi, D. Mazzei, Towards an [35] K. Papineni, S. Roukos, T. Ward, W.-J. Zhu, Bleu: a
llm-based intelligent assistant for industry 5.0, in: method for automatic evaluation of machine
transProceedings of the 1st International Workshop on lation, in: Proceedings of the 40th annual meeting
Designing and Building Hybrid Human–AI Systems of the Association for Computational Linguistics,
(SYNERGY 2024), volume 3701, 2024. URL: https: 2002, pp. 311–318.</p>
      <p>//ceur-ws.org/Vol-3701/paper7.pdf. [36] C.-Y. Lin, Rouge: A package for automatic
eval[23] J. Johnson, M. Douze, H. Jégou, Billion-scale simi- uation of summaries, in: Text summarization
larity search with GPUs, IEEE Transactions on Big branches out, 2004, pp. 74–81.</p>
      <p>Data 7 (2019) 535–547. [37] T. Alkire, C. Rosen, Romance languages: A
histori[24] N. Muennighof, N. Tazi, L. Magne, N. Reimers, cal introduction, Cambridge University Press, 2010.</p>
      <p>Mteb: Massive text embedding benchmark, arXiv
preprint arXiv:2210.07316 (2022). URL: https:
//arxiv.org/abs/2210.07316. doi:10.48550/ARXIV.</p>
      <p>2210.07316.</p>
      <p>
        A. unipa-corpus details
unipa-corpus [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] is a collection of Italian documents
that were retrieved directly from the website of the
University of Palermo in Semptember 2023. The corpus is
divided in two main sections, namely Education, that
groups the available bachelor and master degree courses,
and Future Students where important information about
taxes payment and enrollment procedure are reported.
      </p>
      <p>
        For fine-tuning purposes, a semi-automatic procedure,
involving gpt-3.5-turbo [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], was implemented to build
a QA dataset. In Table 3 are reported the statistics of
unipa-corpus.
      </p>
    </sec>
    <sec id="sec-6">
      <title>B. Inference QA pairs</title>
      <p>Answers
Il professore di Intelligenza Artificiale 1 del corso di Laurea Magistrale
in Ingegneria Informatica è il professore Gaglio e l’insegnamento verrà
erogato durante il primo semestre. Per maggiori informazioni vai su
http://www.unipa.it/struttura.html?id=721
La presentazione delle domande di iscrizione per l’Anno Accademico
2023/2024 varia in base alla tipologia di corso. Per i corsi di studio
triennali e magistrali a ciclo unico, la domanda può essere presentata
dall’1 agosto al 30 settembre 2023 , mentre per i corsi magistrali, dall’1
agosto al 30 novembre 2023, in entrambi i casi, è richiesto il pagamento
della prima rata delle tasse universitarie. Per i corsi di studio ad
accesso programmato locale, con prenotazione online o ad accesso
programmato nazionale, le immatricolazioni vanno efettuate entro i
termini previsti dal relativi bandi di concorso.</p>
      <p>Il corso di laurea magistrale in Chimica presso l’Università degli Studi
di Palermo, ha sede a Palermo e ha una durata di 2 anni. L’obiettivo
del corso è la formazione di laureati con una solida preparazione di
base, che li ponga in grado di afrontare con competenza ed autonomia
sia attività professionali sia quelle di ricerca accademica ed industriale.
Per questo motivo il corso di studio è progettato con lo scopo di
completare la cultura scientifica nelle discipline chimiche fondamentali
e di introdurre conoscenze più avanzate. Quindi lo studente può
completare il proprio piano di studi, attenzionando gli insegnamenti
che meglio si adattano ai propri interessi specifici, come la chimica
dei materiali, dell’ambiente e dei beni culturali, la chimica
supramolecolare e la didattica della chimica. Per maggiori informazioni
vai su http://www.unipa.it/struttura.html?id=766
Ciao! Il corso di Laurea in Consulente Giuridico d’Impresa sembra
proprio fare al caso tuo! Il corso di laurea ha sede a Trapani e ha una
durata di 3 anni e ti fornirà le conoscenze in ambito giuridico ed
economico-aziendalistico. Una volta terminato il percorso di studio
potrai svolgere attività interdisciplinari che richiedono competenze
giuridiche, aziendalistiche e organizzativo-gestionali. Per maggiori
informazioni vai su http://www.unipa.it/struttura.htmlid=1557
È possibile recarsi in segreteria il lunedì, mercoledì e venerdì dalle 10.00
alle 12.00, martedì e giovedì dalle 15.00 alle 17.00 . Puoi prenotare il
tuo turno attraverso la App SolariQ. Per maggiori informazioni vai su
https://www.unipa.it/servizi/segreterie/
Il pagamento delle tasse deve essere efettuato esclusivamente mediante
sistema PAgoPA (Pagamenti della Pubblica Amministrazione). Dopo aver
compilato la pratica online, è possibile pagare direttamente online con
il sistema PAgoPA o stampare il bollettino e pagare presso tabaccai
convenzionati o ricevitorie abilitate PAgoPA. Ulteriori informazioni sul
pagamento via PAgoPA sono reperibili qui https://immaweb.unipa.it/
immaweb/public/pagamenti.seam, mentre è disponibile il Regolamento in
materia di contribuzione studentesca https://www.unipa.it/servizi/segreterie/
.content/documenti/regolamenti_calendari/2023/5105144def_regolamento-contribuzione–studentesca-2023—24-2.pdf</p>
      <p>Answers
The Artificial Intelligence 1 professor for the Computer Engineering
Master degree course is Professor Gaglio and it will be delivered
during the first semester. For more information go to
http://www.unipa.it/struttura.html?id=721
The submission of applications for the Academic Year 2023/2024 varies
according to the type of course. For three-year and single-cycle
master’s degree courses, applications can be submitted from 1 August
to 30 September 2023, while for master’s degree courses, from 1 August
to 30 November 2023; in both cases, payment of the first instalment of
tuition fees is required. For courses with local programmed access,
with online booking or national programmed access, enrolment must be
carried out by the deadlines set out in the corresponding calls for
application.</p>
      <p>The Master’s degree course in Chemistry at the University of Palermo
is based in Palermo and lasts 2 years. The aim of the course is to train
graduates with a good background, enabling them to deal competently
and independently with both professional activities and academic
and industrial research. For this reason, the course is designed
with the aim of completing the scientific culture in the fundamental
chemical disciplines and introducing more advanced knowledge.
Therefore, students can complete their study plan by focusing on the
subjects that best suit their specific interests, such as the chemistry
of materials, the environment and cultural heritage, supramolecular
chemistry and the didactics of chemistry. For more information
go to http://www.unipa.it/struttura.html?id=766
Hi! The Bachelor of Business Law Consultant programme sounds like it
could be just the thing for you! The degree course is based in Trapani
and lasts 3 years and will provide you with knowledge in the fields of
law and business economics. Once you have completed the course you
will be able to carry out interdisciplinary activities requiring legal,
business and organisational-managerial skills. For more information
go to http://www.unipa.it/struttura.html?id=1557
You can go to the secretariat on Mondays, Wednesdays and Fridays from
10 a.m. to 12 noon, Tuesdays and Thursdays from 3 p.m. to 5 p.m. .
You can book your appointment through the SolariQ App. For more
information go to https://www.unipa.it/servizi/segreterie/
Fees must be paid exclusively through the PAgoPA (Public Administration
Payments) system, which is accessed through the university portal. After
completing the paperwork online, you can either pay directly online via
the PAgoPA system or print out the payment slip and pay at a
PAgoPA-enabled tax ofice. Further information on paying via PAgoPA
can be found here https://immaweb.unipa.it/immaweb/public/pagamenti.
seam, while the Student Contribution Regulations is available here
https://www.unipa.it/servizi/segreterie/.content/documents/regulations_
calendars/2023/5105144-def_regulation-student-contribution-2023-24-2.pdf</p>
    </sec>
    <sec id="sec-7">
      <title>C. Embedding spaces</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Albert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almahairi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Babaei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bashlykov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhargava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhosale</surname>
          </string-name>
          , et al.,
          <source>Llama</source>
          <volume>2</volume>
          :
          <article-title>Open foundation and fine-tuned chat models</article-title>
          ,
          <source>arXiv preprint arXiv:2307.09288</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A. . M.</given-names>
            <surname>Llama Team</surname>
          </string-name>
          ,
          <source>The llama 3 herd of models</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2407.21783. arXiv:
          <volume>2407</volume>
          .
          <fpage>21783</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          , et al.,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4] OpenAI, Gpt-4
          <source>technical report</source>
          ,
          <year>2024</year>
          . URL: https: //arxiv.org/abs/2303.08774. arXiv:
          <volume>2303</volume>
          .
          <fpage>08774</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xie</surname>
          </string-name>
          , S.-
          <string-name>
            <given-names>Z. J.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Parameter-eficient fine-tuning methods for pretrained language models: A critical review and assessment, 2023</article-title>
          . URL: https://arxiv.org/abs/2312. 12148. arXiv:
          <volume>2312</volume>
          .
          <fpage>12148</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wallis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Allen-Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Lora:
          <article-title>Low-rank adaptation of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2106.09685</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Dettmers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Belkada</surname>
          </string-name>
          , L. Zettlemoyer,
          <string-name>
            <surname>Llm.</surname>
          </string-name>
          int8()
          <article-title>: 8-bit matrix multiplication for transformers at scale</article-title>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/ 2208.07339. arXiv:
          <volume>2208</volume>
          .
          <fpage>07339</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Polignano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Basile</surname>
          </string-name>
          , G. Semeraro,
          <article-title>Advanced natural-based interaction for the italian language: Llamantino-3-anita</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2405</volume>
          .
          <fpage>07101</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Orlando</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Moroni</surname>
          </string-name>
          , P.-L. Huguet
          <string-name>
            <surname>Cabot</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Conia</surname>
            , E. Barba,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Navigli</surname>
          </string-name>
          ,
          <source>Minerva technical report</source>
          ,
          <year>2024</year>
          . URL: https://nlp.uniroma1.it/minerva/.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>I. Siragusa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pirrone</surname>
          </string-name>
          ,
          <article-title>Unipa-gpt: Large language models for university-oriented qa in italian, 2024</article-title>
          . URL: https://arxiv.org/abs/2407.14246. arXiv:
          <volume>2407</volume>
          .
          <fpage>14246</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Küttler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          , W.-t. Yih,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          , et al.,
          <article-title>Retrieval-augmented generation for knowledge-intensive nlp tasks</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>9459</fpage>
          -
          <lpage>9474</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL:
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>