<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">VeryfIT -Benchmark of Fact-Checked Claims for Italian: A CALAMITA Challenge</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jacopo</forename><surname>Gili</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Turin</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Viviana</forename><surname>Patti</surname></persName>
							<email>viviana.patti@unito.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Turin</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lucia</forename><surname>Passaro</surname></persName>
							<email>lucia.passaro@unipi.it</email>
							<affiliation key="aff1">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Pisa</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tommaso</forename><surname>Caselli</surname></persName>
							<email>t.caselli@rug.nl</email>
							<affiliation key="aff2">
								<orgName type="department">CLCG</orgName>
								<orgName type="institution">University of Groningen</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">VeryfIT -Benchmark of Fact-Checked Claims for Italian: A CALAMITA Challenge</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">47793B8231AFD3746C1B6DA2E538AEE6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>fact checking</term>
					<term>benchmark</term>
					<term>factual knowledge</term>
					<term>Italian</term>
					<term>fake news</term>
					<term>CALAMITA</term>
					<term>CheckIT! 1. Challenge: Introduction and Motivation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Achieving factual accuracy is a known pending issue for language models. Their design centered around the interactive component of user interaction and the extensive use of "spontaneous" training data, has made them highly adept at conversational tasks but not fully reliable in terms of factual correctness. VeryfIT addresses this issue by evaluating the in-memory factual knowledge of language models on data written by professional fact-checkers, posing it as a true or false question. Topics of the statements vary but most are in specific domains related to the Italian government, policies, and social issues. The task presents several challenges: extracting statements from segments of speeches, determining appropriate contextual relevance both temporally and factually, and ultimately verifying the accuracy of the statements.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The pollution of the information ecosystem by means of misleading or false information has reached unprecedented levels at a global scale. This has been possible thanks to a combination of multiple factors, among which the collapse of (local and national) journalism; an increasing sense of distrust in science and evidence-based facts; and the presence of computational amplification tools such as bots <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. In this sense the rise of Large Language Models (LLMs) with the constant increase of their performances has introduced both opportunities and challenges in the fight against misinformation: while LLMs possess the capability to generate coherent and contextually relevant text, they also pose risks by potentially producing deceptive misinformation at scale <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>.</p><p>Testing factual and common sense knowledge in LLMs has been a common although not easy task involving mostly multi-choice question answering, a method easy to automate and not prone to ambiguity, and spanning across wide ranges of academic and professional domains like mathematics, medicine, history, law, general knowledge and many others <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12]</ref>.</p><p>Developing benchmarks to test the ability of LLMs to accurately evaluate factual knowledge is more relevant than ever considering the ease of access of these tools to non-experts for any purpose (entertainment, education, professional settings) and the increasing integration of these technologies in every day activities. Notably, most of these tasks and corresponding benchmarks are in English with other languages being represented through machine-translated data or no data at all. This is true for Italian too. For instance, SQUAD-IT <ref type="bibr" target="#b13">[13]</ref> is a machine-translated version of the SQUAD dataset <ref type="bibr" target="#b14">[14]</ref> and it is the reference for evaluating models on QA-tasks.</p><p>While machine-translation has been constantly improving, it can indeed easily introduce artefacts in the output text impairing naturalness and correctness, moreover translated data can be subjected to the loss of nuance and context as translations may not capture cultural nuances or contextual meanings, leading to misunderstandings or misinterpretations in the target language: certain phrases or idioms may not have direct equivalents in other languages, and the presence of linguistic constructions typical of the source language may be encouraged excessively <ref type="bibr" target="#b15">[15]</ref>.</p><p>By using data from a professional fact-checking agency 1 we can test knowledge memorization of LMs and to what extend intra-memory conflicts, resulting in "hallucinations", arise. Furthermore, doing so using Italian data centered around the Italian and European contexts ensures testing LM's functionalities directly in Italian.</p><p>This task is based on CheckIT! <ref type="bibr" target="#b16">[16]</ref>, a resource of expert fact-checked claims designed to fill a gap for the development of AI-assisted fact-checking pipelines for Italian.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Challenge: Description</head><p>The challenge is a binary classification task in a zero-shot setting: for each atomic statement, any LM is asked to determine its factuality with respect to the time it was uttered by answering only with one of the two labels, "Vero" (true) or "Falso" (false). A third label for half true statements could have been easily kept as it was already part of the dataset from which the data is sourced, but in this first stage we opted for the binary setting as to limit task complexity.</p><p>Some cases in the dataset exhibit complexities due to the combination of multiple pieces of information within a single claim, which can affect the final determination of veracity. For instance, consider the following scenario:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Original claim Translation</head><p>«Se è vero che oltre l'82% dei morti da Covid hanno più di 70 anni, non si capisce perché meno della metà degli over 80 sia stato vaccinato finora» «If it is true that over 82% of Covid deaths are over 70 years old, it is not clear why less than half of those over 80 have been vaccinated so far»</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1 Example of a claim</head><p>The informations concerning this statement are:</p><p>1. Out of all the deceased due to the Covid19 pandemic, 82% are people over 70 years old.</p><p>2. Less than half of the citizens over 80 years old had administered at least one dose of vaccine against Covid19.</p><p>This example also highlights the importance of incorporating the appropriate temporal context in the verification process. Factual information, especially involving statistics or reports about the state of the world, evolves over time and failing to account for this can invalidate the conclusions drawn by experts. Although more complex statements require a broader knowledge base, by now language models have shown understanding abilities well over this level and should not be subjugated by it.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data description</head><p>The VeryfIT dataset consists of 2,021 claims taken from CheckIT! <ref type="bibr" target="#b16">[16]</ref>. Not all claims were included due to the binary format of the task as VeryfIT classifies claims as either "Vero" <ref type="bibr">[True]</ref> or "Falso" [False], whereas CheckIT! recognizes an intermediate "Ni" [Half true] label. As a result, all claims with the "half-true" verdict were discarded.</p><p>Furthermore, we considered pertaining to the task to provide also a smaller subset of claims, "VeryfIT_small", balanced on the political orientation of the politician speaking, as misinformation can occur on all topics but when referring to political misinformation each side of the political spectrum has some more widespread topics and recurrent formulations.</p><p>Additionally, an annotation task was carried out on the VeryfIT_small subset aimed at the clarification of statements presenting a level of ambiguity that would have proven detrimental to the task: around 12% of the statements have available an alternative version "enriched" of informations vital to the task. We will refer to them as "enriched statements" (subsection 3.2).</p><p>In conclusion, 2 versions of the dataset are available: VeryfIT (2,021 claims) and VeryfIT_small (352 claims of which 43 with an enriched version).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Creation of VeryfIT_small</head><p>The first step to achieve this goal was to exclude around 400 out of the 2,021 claims of VeryfIT for which information about the political orientation of the speaker was not available.</p><p>We then mapped, using Wikipedia as a source, the political orientation of the parties (and thus of the authors of the claims at the moment of remark) into eight finegrained, commonly recognized political categories: farleft, left, center-left, center, center-right, right, far-right. An illustration on the list of all the parties and their corresponding political orientation is reported in Table <ref type="table" target="#tab_0">2</ref>. An additional label 'transverse' was added to indicate a non precise placement in the political spectrum. This label includes one party ("Movimento 5 Stelle"), members of the Italian institutions above political parties (e.g. the President of the Republic), and experts not affiliated to any political party or political coalition like members of a technical government <ref type="foot" target="#foot_0">2</ref> .</p><p>At first glance, the Italian political spectrum may appear only slightly unbalanced. Despite the absence of a far-left representation, the distribution of parties across the spectrum is relatively symmetrical. Out of the 23 political parties in the data, six are from the left, two from the center-left, six from the center, three from the center-right, two from the right, and three from the farright. However, the distribution of claims is not as well balanced, with a larger number of claims from the rights and far-right parties than the rest as reported in table 3.</p><p>To ensure the balance of our benchmark we decided to reduce the label granularity from eight to four, by col-</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Political party</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Orientation label</head><p>Alleanza   VeryfIT_small: Final distribution of verdict labels in the political spectrum. Highlighted in green the number of labels of enriched statements (explained in subsection 3.2).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Enriched statements</head><p>Given the specificity of the statements, many of which require detailed knowledge of topics related to Italian institutions and policies, and the occasional ambiguity arising from their oral nature, the task has been further divided into two sub-tasks with slight data modifications, aimed at adding vital context to statements that were excessively reliant on information external to the statements themselves. The altered statements account for around 12% of the VeryfIT_small dataset, as excessive human intervention would undermine the core principle of testing on natural data, aligned with what language models might be asked to handle in real-life scenarios. In most cases, minimal adjustments were made, such as retaining the original claim but adding the name of the politician speaking or clarifying specific references.</p><p>The goal of partially or entirely removing the initial layer of complexity, by simplifying the extraction of the relevant information from the statement for verification, is to highlight a stronger correlation between the benchmark results and the language model's actual factual knowledge: when working with natural data, the model's responses may stem from its difficulty in comprehending the specific information it is being asked to verify. However, with altered data, its responses are more directly influenced by gaps in its knowledge.</p><p>Examples </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>L'elezione dei membri della</head><p>Corte Costituzionale e del Consiglio Superiore della Magistratura (Csm) è un dovere che la costituzione italiana dà al parlamento.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 6 Comparison of Original and Enriched Statements</head><p>The reasons for enriching the statements in table 6 all revolve around the lack of pivotal information to determine factuality: The first statement is completely missing the context and presents an unclear term "grandi elettori" [big voters], relatively known in the political context, but that could be mistaken for a physical feature or for a consideration regarding the age of voters; the second statement has an unclear formulation as "pagare" [to pay] does not refer univocally to taxes; the third statement is missing the subject; the fourth and last statement is missing part of its context as "stiamo facendo un lavoro" [we are doing a job] "stiamo svolgendo un ruolo" [we are playing a role] both refer to a very specific duty of the parliament that does not get mentioned directly.</p><p>Preliminary results obtained through the chat function of Claude 3.5 Sonnet<ref type="foot" target="#foot_1">3</ref> and GPT-4o <ref type="foot" target="#foot_2">4</ref> show that respectively two out of the four statements (Claude) and one out of the four statements (GPT) reported in Table <ref type="table">6</ref> get wrongly classified when presented in the original version, while providing the models with the enriched versions brings up the correctly classifications to four out of four for both models. These results however can only partially prove the effectiveness of enriched statements as different models when presented a partial context could provide different verdicts, even guessing the right one.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Annotation details</head><p>During the making of the VeryfIT datasets, it was noticed that not all the statements were actual claims: in articles with multiple claims to check, the 'statement' field was filled with a short title resuming them all, often in the format "[name of the politician] on [topic]". Regular expressions were used to highlight statements not starting with '"' or '«', the two symbols used to denote a dialogue or part of a speech, and a manual check brought to the exclusion of around 170 statements. Another important annotation step has been producing the enriched statements. A human annotator<ref type="foot" target="#foot_3">5</ref> reviewed the VeryfIT_small dataset, identifying statements that could benefit from additional context, and produced enriched variations of those statements. In most cases, minimal adjustments were made, such as retaining the original claim but adding the name of the politician speaking or clarifying anaphoric references.</p><p>The decision of applying this annotation step to the VeryfIT_small subset, instead of the full dataset, is related to the amount of manual work it would have required.</p><p>Additionally another annotation step involved completing the "macro_area" [topic] field for all the 352 entries of VeryfIT_small. Although this field was included in the original dataset, it was missing a value in approximately 15% of the entries. This was done manually, classifying statements into the pre-existing topic labels which are: 'questioni sociali' [social matters], 'economia' [economy], 'esteri' [foreign affairs], 'giustizia' [justice], 'istituzioni' [institutions], 'ambiente' [environment], 'altro <ref type="bibr">' [others]</ref>. The new labels were chosen by comparing unlabelled statements with statements that already had a label and inspecting the contents of the articles from which they were extracted, sometimes only needing to look at the 'tags' field to find all the information needed. To avoid even the smallest imprecision that would have impaired the original label system made by journalist, non-certain labels were put in the 'altro' category.</p><p>Statistics about the distribution of these labels can be found in section 3.6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Data format</head><p>Brief explanation of the data fields:</p><p>• annotato: If True, the statement has a revised version.</p><p>• id: ID of the corresponding article in CheckIT!.</p><p>• statement_date: Date of statements diffusion.</p><p>• statement: The statement.</p><p>• verdict: Factuality verdict.</p><p>• orientamento: Orientation of the political party of the politician author of the statement.</p><p>• macro_area: Topic of the statement.</p><p>• tags: List of tags.</p><p>• statement_revised: Revised version of the statement, if present.</p><p>Fields such as 'macro_area' and 'tags' serve as indicators of the topic, the former providing a general categorization and the latter offering more specific details. These informations were included with in mind future tasks that could reveal differences in factual knowledge across different subjects.</p><p>{ "annotato": False, "id": 991, "statement_date": 2019-07-12, "statement": "[Il salario minimo n.d.r.] Manca solo a noi e ai Paesi dell'Est Europa", "verdict": "Falso", "orientamento": 'C', "macro_area": "questioni sociali", "tags": "['questioni sociali', 'panzana pazzesca', 'italia', 'eu', 'salario minimo']", "statement_revised": "" }, { "annotato": True, "id": 123, "statement_date": 2023-02-14, "statement": "Il canone in bolletta fu una mia scelta. Costava 113 euro. Averlo fatto pagare a tutti ha portato a un abbassamento del costo da 113 a 90 euro ", "verdict": "Vero", "orientamento": 'C', "macro_area": "altro", "tags": "["canone", "rai", "bolletta", "</p><p>costo"]", "statement_revised": "Il canone in bolletta fu una mia scelta [di Matteo Renzi]. Costava 113 euro. Averlo fatto pagare a tutti ha portato a un abbassamento del costo da 113 a 90 euro" } </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Example of prompts used for zero shots</head><p>The models are expected to be evaluated on this task in a zero-shot setting, thereby also better resembling the conditions of a real use-case scenario. The prompt we suggest to use for the evaluation is basic, and urges the model to limit its answer to just the letter corresponding to the answer. The original prompt in Italian, together with its English translation, are reported in Box 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt</head><p>Il seguente statement, nella data indicata, è vero o falso? Rispondi solo con "Vero" o "Falso".</p><p>The following statement, on the date indicated, is true or false? Answer only with "True" or "False".</p><p>Box 1: Zero-shot prompt</p><p>The prompt does not contain any information about the subject of the question or any other informative cues apart from the time reference needed to anchor the claim in a temporal context. In this way, our benchmark not only tests the model in question answering, but also indirectly tests the instruction-following abilities of the model in a language different than English.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">Detailed data statistics</head><p>The full VeryfIT! dataset is composed of 2,021 entries in the italian language. Out of these claims, 352 form the VeryfIT_small dataset in which the entries are equally split across the three main sides of a semplification of the classical political spectrum (left, right, center) and a fourth label 'trasversal', used to address non precise placement in the political spectrum or complete absence of affiliation to any political party or political coalition.</p><p>Of the 352 claims in the VeryfIT_small dataset, 43 have available an enriched variation of the statement, providing additional context alongside the original statement.</p><p>The distribution of claims and factuality labels across topics is presented in Table <ref type="table" target="#tab_6">8</ref>, Table <ref type="table" target="#tab_8">9</ref>, Table <ref type="table" target="#tab_9">10</ref>  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Metrics</head><p>Accuracy serves as the evaluation metric of the task due to its intuitive interpretation and broad applicability. Accuracy provides a clear measure of a classifier's overall performance by calculating the proportion of correct predictions among total cases examined.</p><p>No other metrics were chosen for the task.   total 88 <ref type="bibr" target="#b15">[15]</ref> 88 <ref type="bibr" target="#b10">[11]</ref> 88 <ref type="bibr" target="#b8">[9]</ref> 88 <ref type="bibr" target="#b7">[8]</ref> Table <ref type="table">11</ref> VeryfIT_small: Distribution of claims per topic and positioning in the simplified political spectrum. Highlighted in green the number of labels of enriched statements.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Limitations</head><p>The totality of the data comes from an expert, reliable source. For this reason, the quality of the verdicts is assured to be high. One possible limitation is due to the time-relatedness of said verdicts: claims can be truth and false at times depending on the temporal context in which they are evaluated. LMs could have an hard time discerning informations pertaining specific time intervals, given that they could also not have been trained on data related to them. Another limitation could be the depth of the factual knowledge required to understand and consequently answer the questions of the dataset. As previously stated, VeryfIT data is about italian/european context and touches details of various fields that most probably not even the citizens would know about! Remarkably, the risk of the data being present in training corpuses for LMs should be mitigated as the CheckIT! dataset is not publicly released.</p><p>Finally, fact-checking is a very complex task and statements could carry different degrees of truthness, more than a binary setting can express. We chose to limit for now the task to a binary classification challenge to not make it too complicated, but we do not exclude further development towards a multi-label setting to better capture the nuances of the fact-checking process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Ethical issues</head><p>No ethical issue has arisen from the making of this task, all the data has been sourced through agreements with the original authors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Data license and copyright issues</head><p>The data cannot be publicly released due to a Data Sharing Agreement between University of Groningen and Pagella Politica. At the moment of writing of this contribution to obtain VeryfIT! contact dr. Tommaso Caselli.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Data format</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="9,129.94,117.78,333.35,249.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 2</head><label>2</label><figDesc>VeryfIT data: Italian political parties and their orientation.</figDesc><table><row><cell>Verdi e Sinistra</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Alternativa Popolare</cell><cell cols="2">center-right</cell><cell></cell></row><row><cell>Articolo Uno</cell><cell cols="2">center-left</cell><cell></cell></row><row><cell>Azione</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Coraggio Italia</cell><cell cols="2">center-right</cell><cell></cell></row><row><cell>Europa Verde</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Forza Italia</cell><cell>right</cell><cell></cell><cell></cell></row><row><cell>Fratelli d'Italia</cell><cell>far-right</cell><cell></cell><cell></cell></row><row><cell>Impegno Civico</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Indipendente</cell><cell cols="2">transverse</cell><cell></cell></row><row><cell>Italexit</cell><cell>far-right</cell><cell></cell><cell></cell></row><row><cell>Italia Viva</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Lega Nord</cell><cell>far-right</cell><cell></cell><cell></cell></row><row><cell>Liberi e uguali</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Movimento 5 Stelle</cell><cell cols="2">transverse</cell><cell></cell></row><row><cell>Nuovo Centro Destra</cell><cell cols="2">center-right</cell><cell></cell></row><row><cell>Partito Democratico</cell><cell cols="2">center-left</cell><cell></cell></row><row><cell>Più Europa</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Popolo della Libertà</cell><cell>right</cell><cell></cell><cell></cell></row><row><cell>Possibile</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Radicali Italiani</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Scelta Civica</cell><cell>center</cell><cell></cell><cell></cell></row><row><cell>Sinistra Ecologia Libertà</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Sinistra italiana</cell><cell>left</cell><cell></cell><cell></cell></row><row><cell>Tecnico</cell><cell cols="2">transverse</cell><cell></cell></row><row><cell></cell><cell cols="2">Claims</cell><cell></cell></row><row><cell cols="4">Political side True False Total</cell></row><row><cell>Left</cell><cell>44</cell><cell>28</cell><cell>72</cell></row><row><cell>Center-left</cell><cell>323</cell><cell>110</cell><cell>433</cell></row><row><cell>Center</cell><cell>105</cell><cell>82</cell><cell>187</cell></row><row><cell>Center-right</cell><cell>8</cell><cell>2</cell><cell>10</cell></row><row><cell>Right</cell><cell>79</cell><cell>84</cell><cell>163</cell></row><row><cell>Far-right</cell><cell>156</cell><cell>241</cell><cell>397</cell></row><row><cell>Transverse</cell><cell>209</cell><cell>146</cell><cell>355</cell></row><row><cell>total</cell><cell>924</cell><cell>693</cell><cell>1,617</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 3</head><label>3</label><figDesc>VeryfIT data after exclusion of claims where information about political orientation of the speaker was not available: Distribution of verdict labels in the political spectrum.</figDesc><table><row><cell>The re-aggregated coarse-grained labels are reported in</cell></row><row><cell>Table 4.</cell></row><row><cell>Although the distribution is still unbalanced between</cell></row></table><note>lapsing labels far-left, left and center-left into 'left' [SX], and far-right, right and center-right into 'right' [DX]. Labels center [C] and trasversal [T] remained untouched.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 5</head><label>5</label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head></head><label></label><figDesc>of enriched statements are reported in Table 6:</figDesc><table><row><cell>Original statement</cell><cell>Enriched statement</cell></row><row><cell>Abbiamo 490 grandi elet-</cell><cell>Gli elettori dell'area di cen-</cell></row><row><cell>tori</cell><cell>trosinistra che voteranno</cell></row><row><cell></cell><cell>per l'elezione del Presi-</cell></row><row><cell></cell><cell>dente della Repubblica</cell></row><row><cell></cell><cell>saranno 490.</cell></row><row><cell>Oggi in Italia sono 796</cell><cell>Oggi in Italia sono 796</cell></row><row><cell>quelli che pagano più di 1</cell><cell>quelli che dichiarano un</cell></row><row><cell>milione di euro</cell><cell>reddito superiore ad 1 mil-</cell></row><row><cell></cell><cell>ione di euro.</cell></row><row><cell>[Alle europee] io ho battuto</cell><cell>[Alle europee] io [Carlo</cell></row><row><cell>Salvini in molti capoluoghi</cell><cell>Calenda] ho battuto</cell></row><row><cell>di provincia</cell><cell>Salvini in molti capoluoghi</cell></row><row><cell></cell><cell>di provincia.</cell></row><row><cell>In parlamento stiamo</cell><cell></cell></row><row><cell>facendo un lavoro che</cell><cell></cell></row><row><cell>risponde a una prerogativa</cell><cell></cell></row><row><cell>costituzionale. Certamente</cell><cell></cell></row><row><cell>si sarebbero tutti auspicati,</cell><cell></cell></row><row><cell>me compresa, tempi più</cell><cell></cell></row><row><cell>brevi ma non stiamo</cell><cell></cell></row><row><cell>perdendo tempo. Stiamo</cell><cell></cell></row><row><cell>svolgendo un ruolo che ci</cell><cell></cell></row><row><cell>compete e che la Costi-</cell><cell></cell></row><row><cell>tuzione da' al parlamento.</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 7</head><label>7</label><figDesc>Moreover around 30 statements with formats resembling "[name of the politician] is [right/wrong] on [topic]: [statement]" were reformulated as claims by removing hints about the factuality verdict and the author of the statement. A couple examples are brought up in table 7. Examples of reworded statements</figDesc><table><row><cell>Original statement</cell><cell>Reworded statement</cell></row><row><cell>Giulia Grillo sbaglia: i</cell><cell>i medici e gli infermieri ital-</cell></row><row><cell>medici e gli infermieri ital-</cell><cell>iani sono i meno pagati</cell></row><row><cell>iani non sono i meno pagati</cell><cell></cell></row><row><cell>Secondo Di Maio il governo</cell><cell>Il governo investe nelle cen-</cell></row><row><cell>investe nelle centrali a car-</cell><cell>trali a carbone</cell></row><row><cell>bone, ma è il contrario</cell><cell></cell></row><row><cell>No, per la Corte dei Conti</cell><cell>Per la Corte dei Conti</cell></row><row><cell>non ci saranno 17 miliardi</cell><cell>ci saranno 17 miliardi di</cell></row><row><cell>di nuove tasse</cell><cell>nuove tasse</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head></head><label></label><figDesc>, Table11.</figDesc><table><row><cell></cell><cell></cell><cell>Claims</cell><cell></cell></row><row><cell>Macro_area</cell><cell cols="3">True False Total</cell></row><row><cell>questioni sociali</cell><cell>256</cell><cell>170</cell><cell>426</cell></row><row><cell>economia</cell><cell>264</cell><cell>155</cell><cell>419</cell></row><row><cell>istituzioni</cell><cell>243</cell><cell>77</cell><cell>320</cell></row><row><cell>esteri</cell><cell>105</cell><cell>53</cell><cell>158</cell></row><row><cell>giustizia</cell><cell>60</cell><cell>26</cell><cell>86</cell></row><row><cell>altro</cell><cell>46</cell><cell>32</cell><cell>78</cell></row><row><cell>ambiente</cell><cell>42</cell><cell>18</cell><cell>60</cell></row><row><cell>un-noted</cell><cell>180</cell><cell>294</cell><cell>474</cell></row><row><cell>total</cell><cell>1,196</cell><cell>825</cell><cell>2,021</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 8</head><label>8</label><figDesc>VeryfIT: Distribution of claims and factuality labels per topics ordered by total value.Further statistics on the original CheckIT! dataset is available in Figure A andTable A in Appendix A.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 9</head><label>9</label><figDesc>VeryfIT data after exclusion of claims where information about political orientation of the speaker was not available: Distribution of claims per topic and positioning in the political spectrum.</figDesc><table><row><cell></cell><cell></cell><cell>Claims</cell><cell></cell></row><row><cell>Macro_area</cell><cell>True</cell><cell>False</cell><cell>Total</cell></row><row><cell>questioni sociali</cell><cell>50 [2]</cell><cell>37 [4]</cell><cell>87 [6]</cell></row><row><cell>economia</cell><cell>53 [4]</cell><cell>37 [4]</cell><cell>90 [8]</cell></row><row><cell>istituzioni</cell><cell>46 [11]</cell><cell>17 [5]</cell><cell>63 [16]</cell></row><row><cell>esteri</cell><cell>26 [4]</cell><cell>19 [3]</cell><cell>45 [7]</cell></row><row><cell>ambiente</cell><cell>8 [1]</cell><cell>10</cell><cell>18 [1]</cell></row><row><cell>giustizia</cell><cell>7</cell><cell>8</cell><cell>15</cell></row><row><cell>altro</cell><cell>10 [1]</cell><cell>24 [4]</cell><cell>34 [5]</cell></row><row><cell>total</cell><cell cols="3">200 [23] 152 [20] 352 [43]</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 10</head><label>10</label><figDesc>VeryfIT_small: Distribution of claims and factuality labels per topics ordered by total value. Highlighted in green the number of labels of enriched statements.</figDesc><table><row><cell></cell><cell></cell><cell cols="2">Orientation label</cell><cell></cell></row><row><cell>Macro_area</cell><cell>SX</cell><cell>C</cell><cell>DX</cell><cell>T</cell></row><row><cell>questioni sociali</cell><cell>22 [2]</cell><cell cols="3">15 [1] 25 [2] 25 [1]</cell></row><row><cell>economia</cell><cell>28 [2]</cell><cell cols="2">30 [5] 19 [1]</cell><cell>13</cell></row><row><cell>istituzioni</cell><cell>19 [8]</cell><cell cols="3">8 [1] 15 [3] 21 [4]</cell></row><row><cell>esteri</cell><cell>9 [2]</cell><cell cols="3">13 [1] 12 [2] 11 [2]</cell></row><row><cell>ambiente</cell><cell>2</cell><cell>7</cell><cell>3 [1]</cell><cell>6</cell></row><row><cell>giustizia</cell><cell>2</cell><cell>4</cell><cell>3</cell><cell>6</cell></row><row><cell>altro</cell><cell>6 [1]</cell><cell>11 [3]</cell><cell>11</cell><cell>6 [1]</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://en.wikipedia.org/wiki/Technocratic_government_(Italy)</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://claude.ai/chat</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://chatgpt.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">All the annotations noted in the report was done by the first author of the paper, master student in Computer Science with a background in Natural Language Processing</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Disinformation is on the rise. how does it work?</title>
		<author>
			<persName><forename type="first">T</forename><surname>Economist</surname></persName>
		</author>
		<ptr target="https://www.economist.com/science-and-technology/2024/05/01/disinformation-is-on-the-rise-how-does-it-work" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Wardle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Derakhshan</surname></persName>
		</author>
		<title level="m">Information disorder: Toward an interdisciplinary framework for research and policymaking</title>
				<meeting><address><addrLine>Strasbourg</addrLine></address></meeting>
		<imprint>
			<publisher>Council of Europe</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">27</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<title level="m">Disrupting deceptive uses of ai by covert influence operations</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shu</surname></persName>
		</author>
		<idno type="DOI">10.1002/aaai.12188</idno>
		<ptr target="https://doi.org/10.1002/aaai.12188.doi:10.1002/aaai.12188" />
	</analytic>
	<monogr>
		<title level="m">Combating misinformation in the age of llms: Opportunities and challenges</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Measuring massive multitask language understanding</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hendrycks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Burns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Basart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mazeika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Steinhardt</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2009.03300.arXiv:2009.03300" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Beyond the imitation game: Quantifying and extrapolating the capabilities of language models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Srivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kleyjo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions on Machine Learning Research</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Benchmarking large language models on cmexam-a comprehensive chinese medical exam dataset</title>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>You</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">TruthfulQA: Measuring how models mimic human falsehoods</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Evans</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.229</idno>
		<ptr target="https://aclanthology.org/2022.acl-long.229.doi:10.18653/v1/2022.acl-long.229" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="3214" to="3252" />
		</imprint>
	</monogr>
	<note>: Long Papers), Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Pinto: Faithful language reasoning using promptgenerated rationales</title>
		<author>
			<persName><forename type="first">P</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ilievski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop on Trustworthy and Socially Responsible Machine Learning</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">What disease does this patient have? a large-scale open domain question answering dataset from medical exams</title>
		<author>
			<persName><forename type="first">D</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Oufattole</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-H</forename><surname>Weng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Szolovits</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2009.13081.arXiv:2009.13081" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Commonsenseqa: A question answering challenge targeting commonsense knowledge</title>
		<author>
			<persName><forename type="first">A</forename><surname>Talmor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Herzig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lourie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Berant</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1811.00937.arXiv:1811.00937" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">In-context annotation of topic-oriented datasets of fake news: A case study on the notre-dame fire event</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Passaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondielli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dell'oglio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Marcelloni</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ins.2022.07.128</idno>
		<ptr target="https:" />
	</analytic>
	<monogr>
		<title level="j">Information Sciences</title>
		<imprint>
			<biblScope unit="volume">615</biblScope>
			<biblScope unit="page" from="657" to="677" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title/>
		<idno type="DOI">10.1016/j.ins.2022.07.128</idno>
		<idno>.07.128</idno>
		<ptr target="//doi.org/10.1016/j.ins.2022" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Neural learning for question answering in italian</title>
		<author>
			<persName><forename type="first">D</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zelenanska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Basili</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AI*IA 2018 -Advances in Artificial Intelligence</title>
				<editor>
			<persName><forename type="first">C</forename><surname>Ghidini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Magnini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Passerini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Traverso</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="389" to="402" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Rajpurkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lopyrev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1606.05250.arXiv:1606.05250" />
		<title level="m">Squad: 100,000+ questions for machine comprehension of text</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Melero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Del Pozo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Conde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Reviriego</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayor-Rocher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grandury</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2406.17789</idno>
		<title level="m">Spanish and llm benchmarks: is mmlu lost in translation?</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Checkit!: A corpus of expert fact-checked claims for italian</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gili</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Passaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th Italian Conference on Computational Linguistics, CEUR Workshop Proceedings, CEUR Workshop Proceedings</title>
				<editor>
			<persName><forename type="first">F</forename><surname>Boschetti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Lebani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Magnini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Novielli</surname></persName>
		</editor>
		<meeting>the 9th Italian Conference on Computational Linguistics, CEUR Workshop Proceedings, CEUR Workshop Proceedings</meeting>
		<imprint>
			<publisher>Publisher Copyright</publisher>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="2" to="12" />
		</imprint>
	</monogr>
	<note>9th Italian Conference on Computational Linguistics, CLiC-it 2023 ; Conference date</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
