<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">GATTINA -GenerAtion of TiTles for Italian News Articles: A CALAMITA Challenge</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Maria</forename><surname>Francis</surname></persName>
							<email>maria.francis@unitn.it</email>
							<affiliation key="aff0">
								<orgName type="laboratory">CLCG</orgName>
								<orgName type="institution">University of Groningen</orgName>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Trento</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Matteo</forename><surname>Rinaldi</surname></persName>
							<email>matteo.rinaldi@unito.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jacopo</forename><surname>Gili</surname></persName>
							<email>jacopo.gili584@edu.unito.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Leonardo</forename><surname>De Cosmo</surname></persName>
							<email>leodecosmo@gmail.com</email>
						</author>
						<author>
							<persName><forename type="first">Sandro</forename><surname>Iannaccone</surname></persName>
							<email>iannaccone@galileonet.it</email>
						</author>
						<author>
							<persName><forename type="first">Malvina</forename><surname>Nissim</surname></persName>
							<email>m.nissim@rug.nl</email>
							<affiliation key="aff0">
								<orgName type="laboratory">CLCG</orgName>
								<orgName type="institution">University of Groningen</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Viviana</forename><surname>Patti</surname></persName>
							<email>viviana.patti@unito.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><surname>Galileo</surname></persName>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="department">Tenth Italian Conference on Computational Linguistics</orgName>
								<address>
									<addrLine>Dec 04 -06</addrLine>
									<postCode>2024</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">GATTINA -GenerAtion of TiTles for Italian News Articles: A CALAMITA Challenge</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F057F2220E3E51FAB7F4B2E022444613</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:33+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>CALAMITA Challenge</term>
					<term>Italian</term>
					<term>Benchmarking</term>
					<term>Headline generation</term>
					<term>Summarisation</term>
					<term>LLMs</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We introduce a new benchmark designed to evaluate the ability of Large Language Models (LLMs) to generate Italian-language headlines for science news articles. The benchmark is based on a large dataset of science news articles obtained from Ansa Scienza and Galileo, two important Italian media outlets. Effective headline generation requires more than summarizing article content; headlines must also be informative, engaging, and suitable for the topic and target audience, making automatic evaluation particularly challenging. To address this, we propose two novel transformer-based metrics to assess headline quality. We aim for this benchmark to support the evaluation of Italian LLMs and to foster the development of tools to assist in editorial workflows.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction and Motivation</head><p>The title is undoubtedly one of the most important and crucial components of a journalistic article. A good title intrigues the reader, synthesises the news without anticipating its details, encourages further reading, and is simultaneously pleasant to read or hear. Often, the fate of an article is inextricably linked to the quality of its accompanying title: it is not uncommon for inherently interesting, in-depth, and factually correct articles to go unnoticed simply because they are accompanied by an inappropriate or unattractive title. Composing adequate titles is not a simple operation; it requires experience, sensitivity, balance, a sense of measure, and a deep understanding of the readers. There are no precise and inescapable "rules" -save, of course, for the usual deontological norms of pertinence and truth that regulate the journalistic profession -but in fact, the operation depends almost exclusively on the author's expertise and must be evaluated on a case-by-case basis.</p><p>Factors that can influence the composition of a title include, for example, the topic and the "tone of voice" of the article (a piece reporting a crime news story, for instance, requires a measured, discreet, and respectful title; conversely, a piece on lifestyle can and should be paired with a lighter, ironic, and more colorful title); the style of the publication hosting the article; the destination format (the same article printed in a paper newspaper and published on an online outlet, for example, typically has two different titles); potential "conflicts" with other titles present on the same page (for instance: repetitions of the same word or phrase, or the enunciation of contradictory concepts); space limitations; prescriptions related to search engine optimisation (for example, the use of a particular word or expression particularly popular at the time of publication, or a specific position of words within the title).</p><p>It is in this context that the journalist's toolkit has recently been enriched with a powerful new tool: Large language models (LLMs) undoubtedly have an important role in the world of journalism, including quality journalism. Although incapable of "understanding" content as a human journalist would, as well as the meaning of words, LLMs are naturally capable of producing fluent, complex, plausible, and credible texts in a matter of moments. These models not only can improve the efficiency of editorial processes but also offer new creative and innovative possibilities for content creation, including the automatic generation of journalistic headlines. Analysing why it may be useful for journalism to have an LLM capable of generating titles leads us to consider numerous factors, such as time optimisation, content personalization, and the ability to maintain a high level of quality, coherence, and communicative impact. However, these tools also present many limitations and some dangers, particularly the risk of blindly relying on them.</p><p>Timing and speed, in particular, are one of the great challenges of journalism -being the first to publish a story, especially online, is often essential to attract readers -however, as we have seen, generating effective and incisive titles requires skill and time, which is not always available. An LLM can drastically reduce the time needed to create appropriate titles, for example by suggesting to the author a series of reasoned choices or proposing modifications and corrections to an already written title, always keeping in mind preset criteria such as length, tone, attractiveness, clarity, and the publication's style. Furthermore, if trained on the corpus of a particular publication, an LLM can suggest titles consistent with its tone of voice and editorial history.</p><p>Another important advantage that the use of LLMs can offer is the ability to personalise content for different platforms and audiences. In today's newsrooms, journalists no longer have to worry only about print media but must also consider the web, social media, newsletters, and other digital distribution platforms. Each platform requires a different type of language, style, and length for titles. For example, a title optimised for Twitter (or X) must be short and incisive, while a title for a news website can be more descriptive. An LLM is capable of generating variants of a title based on the medium of dissemination, allowing newsrooms to adapt their content precisely and in a targeted manner. Moreover, using reader behavioural data, the LLM can generate more attractive titles for specific demographic groups, thus improving the engagement and communicative effectiveness of the news.</p><p>With this task, which is developed in the context of the CALAMITA Challenge <ref type="bibr" target="#b0">[1]</ref> and which consists in asking an LLM to generate a headline given the corresponding full article, we have a twofold aim.</p><p>The first aim is to test and analyse the ability of existing and future LLMs on the task of headline generation in the context of Italian news articles. This would provide a substantial step forward compared to past experiments on headline generation for Italian, which were run training much smaller sequence-to-sequence models from scratch <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>. We expect that some of the shortcomings of the automatically generated headlines which were observed in previous work, such as lack of fluency and creativity <ref type="bibr" target="#b1">[2]</ref>, might not affect LLM-based generations.</p><p>The second aim is to provide a reliable, high quality dataset of articles and corresponding headlines in Italian, developed through a direct collaboration of language technology experts and journalists, which can be used and analysed well beyond the CALAMITA challenge. Although similar datasets exist for other languages <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref>, this resource is still lacking for Italian.</p><p>Overall, experimenting with the use of LLMs for title generation can also be considered a first step towards the introduction of more extensive and comprehensive artificial intelligence agents, which assist the journalist in all phases of the creative process, from news research to drafting an outline, to writing the actual piece, and finally to its promotion. Indeed, a close interaction of language models and humans in this task has recently been shown to be key <ref type="bibr" target="#b5">[6]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Challenge Description</head><p>The task of headline generation has often been treated as equal to an extreme summarization task <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b6">7]</ref>. However, simply synthesising the content of the article into a brief description is not enough to provide a satisfying title. Additional characteristics such as attractiveness, creativeness, and many others also play a role. Writing appropriate headlines is challenging, even for current state-of-the-art LLMs.</p><p>Evaluating LLMs on the task of headline generation for Italian news articles thus serves multiple purposes. On one hand, it tests models' capacity to properly understand, that is, to reprocess large source texts in a way that is faithful to the content of the text. On the other hand, it acts as a means to assess the performance of LLMs in many complex dimensions, such as attractiveness, creativity, or adherence to tone. Finally, this benchmark could prove useful in practical applications. For instance, it may help guide decisions on whether, and to what extent, a journal should integrate LLMs into its workflow. It may also serve as an effective testbed for future research and development towards effective deployment in real-world scenarios -One such venue could be the use of prompting to achieve the desired style and tone in generated headlines.</p><p>In our challenge, language models are tasked with generating Italian-language headlines based on articles from scientific news journals written in Italian. Our dataset includes original articles from such journals, along with their human-authored titles. Models are provided the complete source text in the prompt, as well as instructions to generate a title that is brief, coherent, and captivating. We guide the model towards the specific editorial style of the media outlet by including a small number of examples of headlines in our prompt. We employ automatic metrics that assess the model's performance along three dimensions:</p><p>1. Coherency with the original article (HA classifier) 2. Alignment with the style of human written headlines (NS classifier) 3. Similarity between the generated and the goldstandard headline (ROUGE <ref type="bibr" target="#b7">[8]</ref>, SBERT <ref type="bibr" target="#b8">[9]</ref>) However, considering the complexity of the task, we believe that manually reviewing a sample of the generated headlines can offer additional perspectives on the behaviour of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data description</head><p>Our benchmark is based of two datasets consisting of science news articles from two different sources. In each dataset, we provide the full text of the article paired with the original, human-authored headline. Additionally, we include metadata such as link, date, author (if present) and subtitle.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Origin of data</head><p>The data were obtained via web scraping with custom Python scripts. Since links to articles more than a few weeks old are inaccessible on the Ansa website, we collected a large number by downloading the archived "Ansa Scienza" RSS feeds from The Wayback Machine and processing them to remove duplicates and extact links.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Data format</head><p>The data from web scraping were saved in "JSON Lines" (JSONL) format, with each line containing a JSON object with the following fields: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Detailed data statistics</head><p>Our dataset consists of 30,461 articles gathered from two sources:  1. "ANSA scienza", the science section of the Italian newspaper "ANSA", from which obtained 6,889 articles: 649 of which are from 2024, and the others are from a period of time between 2018 and 2022. 2. The "Galileo" website, from which we sourced 23,572 articles dating from April 1996 to May 2024.</p><p>When measured with "tiktoken o200k_base" tokenizer model, we obtained a total of 21,365,897 tokens for the Galileo dataset (average: 906 tokens per article, maximum: 24,306) and a total of 3,762,539 tokens for the Galileo dataset (average: 546 tokens per article, maximum: 7,600). Figures <ref type="figure" target="#fig_2">1 and 2</ref> depict the distribution of articles by token count in the Galileo and Ansa datasets respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Prompting</head><p>Due to the length of each article, the use of task examples in our prompt would be too computationally expensive. Therefore, we test the models in a zero-shot prompting setting. While we do not use any task examples in our prompt, we do provide seven examples of headlines. In this way, the model is given examples of the expected output (a title) rather than examples of the full task (article and title). Professional journalists made a list of 22 headlines that, in their opinion, were representative of a well-made writing process under the three aspects of being captivating, short and informative.</p><p>Each time the model is tested, 7 randomly chosen titles from the list are appended to the standard prompt. As a reference, the identifier of the example headlines is also saved along with the output of the model. See Box 1 for our input prompt.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt for the LLM</head><p>Il tuo compito è generare un titolo accattivante e informativo per l'articolo fornito. Requisiti: -Titolo breve -Cattura l'essenza dell'articolo -Usa un linguaggio vivido e coinvolgente -Non generare alcun tipo di testo che non sia il titolo dell'articolo -Usa esclusivamente l'Italiano. Presta particolare attenzione ai seguenti titoli di esempio e adotta lo stesso stile: Title 1 Title 2 ... Title 7</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Your task is to generate a catchy and informative title for the article provided. Requirements: -Short title -Capture the essence of the article -Use vivid and engaging language -Do not generate any type of text other than the title of the article -Use Italian exclusively. Pay particular attention to the following example titles and adopt the same style: Title 1 Title 2 ... Title 7</head><p>Box 1: Zero-shot prompt and English translation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Preliminary Evaluation</head><p>To get a first impression of LLM performance on our task, we conducted preliminary experiments by manually reviewing headlines generated by several models. Overall, the results were unsatisfactory -while the titles were generally coherent with the articles, they lacked captivation and originality. The majority of the generated headlines followed the format &lt;Keywords: explanation&gt;, leading to repetitive and poorly formulated headlines. Examples of our preliminary results can be found in Table <ref type="table" target="#tab_1">1</ref> in Appendix A. This behaviour persisted even when the models were explicitly instructed to avoid using colons in the titles, or when examples of titles were given. Out of 3,006 headlines generated by Phi-3.5 Mini-Instruct, 2,940 headlines contained a colon. We obtained similar results using Mistral-7B-Instruct-v0.3, Qwen2-7B-Instruct, gemma-2-9b-it and Italia-9B-Instruct-v0.1. Manual experimentation with the commercial LLMs Claude 3.5 Sonnet<ref type="foot" target="#foot_0">1</ref> and ChatGPT 4o<ref type="foot" target="#foot_1">2</ref> yielded the same behaviour:</p><p>• Titolo originale:</p><p>Una rapina cosmica nell'ammasso di galassie dell'Idra • Claude: Rapina cosmica: il furto di gas nell'ammasso dell'Idra • ChatGPT: Rapina Cosmica: NGC 3312 Derubata di Gas nell'Ammasso di Galassie dell'Idra Interestingly, when we asked Claude 3.5 Sonnet to improve our prompt for generating headlines, it added the line &lt;Struttura: [Frase d'impatto o dato interessante]: [Spiegazione o contesto]&gt; to our example prompt, explicitly requesting the unwanted behaviour. It appears that LLMs consistently regard this particular structure as the ideal format for a headline.</p><p>Given the inherent difficulty of interpreting LLM behaviour, we cannot provide a single reason for their preference for this particular construction. Of course, there might be a large presence of such headlines in the training data, particularly from lower-quality journals. There may also be an influence of Search Engine Optimizations (SEO) on the behaviour of the model: Giving importance to keywords is a classic SEO technique.</p><p>Moreover, we generally noticed a preference toward sentences poor in determinative and indefinite articles when compared with human written headlines.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Metrics</head><p>Automatically evaluating the quality of generated headlines is a challenging matter because headline quality is inherently subjective, multi-faceted, and contextdependent. Thus, instead of providing a single numeric value as an overall quality score, headlines should be evaluated along multiple dimensions and subsequently rated for their quality based on specific use cases. To give examples of what others have done -Cafagna et al. <ref type="bibr" target="#b1">[2]</ref> evaluate generated headlines based on the criteria such as grammatical correctness, topic relevance, attractiveness, and overall appropriateness. Cai et al. <ref type="bibr" target="#b9">[10]</ref> assess factors such as factual consistency, relevance, and surface overlap between the generated headline and the article, as well as its alignment with user-specific preferences.</p><p>In the aforementioned papers, the headlines were scored by human evaluators. This approach is resource intensive -to account for differences in individual preferences, hiring multiple human evaluators from varying demographic backgrounds is preferred. This does not scale well to the evaluation of multiple models on largescale benchmarks across multiple studies, making the ability to automatically evaluate the outputs of LLMs essential.</p><p>Historically, n-gram overlap metrics like BLEU <ref type="bibr" target="#b10">[11]</ref>, ROUGE <ref type="bibr" target="#b7">[8]</ref>, or METEOR <ref type="bibr" target="#b11">[12]</ref> have been used to compare generated outputs with reference "gold standard" texts, but these metrics emphasise surface-level matching and are therefore not robust to paraphrasing or other variations in acceptable outputs. Learned metrics such as COMET <ref type="bibr" target="#b12">[13]</ref>, a metric designed to mimic human quality judgement for machine translations, have been gaining in popularity. These are not easily transferable to other languages or tasks, and learnable metrics designed specifically for Italian headline generation are not available. Additionally, such metrics typically produce a single numerical score of 'quality'. To improve interpretability and ensure contextual flexibility, we would prefer to provide individual scores for each dimension. We train two novel learned metrics for Italian headline generation, but leave others for future work.</p><p>We evaluate model performance on our benchmark using four metrics: ROUGE <ref type="bibr" target="#b7">[8]</ref>, SBERT <ref type="bibr" target="#b8">[9]</ref>, and two custom metrics -the Headline-Article and Natural-Synthetic classifiers. Within the context of the CALAMITA challenge, the model's final score will be an aggregate in which four all metrics are weighted equally. Each metric is detailed in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">ROUGE</head><p>ROUGE (Recall-Oriented Understudy for Gisting Evaluation) <ref type="bibr" target="#b7">[8]</ref> is a popular metric used to evaluate automatically generated summarizations. It provides a measure of overlap between generated text and gold-standard references. ROUGE is easily interpretable and allows for easy comparison across many papers due to its widespread use. However, it is not robust to variations in input, making it less suitable for the assessment of tasks involving creativity, such as headline generation. Following others <ref type="bibr" target="#b13">[14]</ref>, we will evaluate our system outputs using ROUGE-L, which identifies the length of the longest common subsequence between system and reference.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">SBERT</head><p>Sentence-BERT, or SBERT <ref type="bibr" target="#b8">[9]</ref>, is a modification of the BERT network that uses Siamese networks and that can derive semantically meaningful, fixed-size vector embeddings from whole sentences. We use SBERT to compare our generated headlines to the gold-standard ones by comparing their SBERT embeddings using cosine similarity, which we then use directly as the similarity score. SBERT produces more meaningful sentence embeddings compared to BERT, which is not designed for sentence similarity tasks -therefore, cosine similarity with BERT embeddings could produce unwanted and less interpretable results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Custom metrics</head><p>Given the limitations of the current available metrics for the headlines generation task, we develop two custom metrics employing classifiers based on Transformer <ref type="bibr" target="#b14">[15]</ref> models. We trained both classifiers on a subset of the "blogs" section of the "Testimole"<ref type="foot" target="#foot_2">3</ref> dataset, which was obtained by web scraping various Italian media sources. Our subset consists of only those parts of the dataset scraped from professional media outlets. The criteria for the selection process, as well as the technical details for each classifier, are in Appendix B.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.1.">HA Classifier</head><p>Our first classifier is based on the Sentence Transformers <ref type="bibr" target="#b8">[9]</ref> architecture, fine-tuned to discriminate between coherent and non-coherent pairs of headlines and articles. A generated headline can score between 0 and 1, representative of the degree of alignment between the headline and the content of the article. Following the work by De Mattei et al. <ref type="bibr" target="#b2">[3]</ref>, we call this classifier "HA", or Headline-Article.</p><p>To train the model, we used a non-finetuned Italian Sentence Bert model <ref type="foot" target="#foot_3">4</ref> to compute an embedding for each article. We then find the headline of the article in the dataset with the highest cosine similarity, and create a new dataset where each row contains the article (anchor), the original title (positive), and the title of the most similar article (negative). Because the original dataset contained some duplicate items, we filtered all articles with "1" as the cosine similarity score. With this dataset, we were able to use Triplet Loss to train the classifier to differentiate between coherent and incoherent titles, starting from the assumption that the original title is the one most coherent with the article's content. We decided to perform a cosine similarity search instead of random shuffling in order to increase the difficulty of the discriminator's task.</p><p>The drawback of this approach is the low context window of the model -all articles were truncated after the first 512 tokens. While it is possible to develop a more complex architecture to account for larger texts, we leave this for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.2.">NS Classifier</head><p>Our second classifier is called "NS", or Natural-Synthetic. It is a binary regression classifier based on an Italian BERT-base uncased model 5 , trained to discriminate between human-authored and machine-generated titles. Given a title as input, the classifier outputs a numerical score indicating the likelihood of the title being close to those written by journalists. We believe that similarity to headlines written by journalists may be a useful indicator of the quality and appropriateness of a generated headline.</p><p>Using the same subset of Testimole employed for the "HA" classifier, we generated over 90,000 synthetic headlines using LLMs of up to 9 billion parameters. To avoid overfitting our classifier to the specific probability distribution of a single model, we generated synthetic headlines using different models; this process is detailed in Appendix C, along with details about the number of generated headlines per model. The result is a labelled dataset containing original as well as generated headlines.</p><p>The advantage of employing a "Natural-Synthetic" classifier is that the training objective is coarse, encouraging the classifier to consider a broad range of aspects that may account for the discrepancy of text generated by machines and humans.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Future works</head><p>We see value in future research using classifiers and regressors to assess specific aspects of generated headlines. Such metrics have the potential to capture complex probability distributions over a multitude of dimensions of the data, including dimensions that are not directly interpretable to human observation. For instance, a learned metric that predicts the amount of attention a headline will generated would be highly useful.</p><p>Inspired by Generative Adversarial Networks (GANs), we find the employment of classification-based metrics promising for developing a model specialized in headline generation. A discriminator/generator training system 5 https://huggingface.co/dbmdz/bert-base-italian-xxl-uncased allows us to build a positive feedback loop in which the headline generation system teaches itself to generate good headlines based on the classification of the discriminator. For instance, the model can be trained to 'fool' the NS discriminator as often as possible while the NS discriminator uses the experience to improve at identifying synthetic data, causing both models to improve simultaneously. This method, for instance, should quickly solve the frequent use of the colon in automatically generated headlines outlined in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Limitations</head><p>Our benchmark is limited to articles and headlines from only two journals, which restricts its representativeness across journalistic domains. As a result, it may not capture the variability present in publications targeting different demographics, covering varied topics, or representing a full spectrum of political perspectives.</p><p>In training our classifiers, we took care to prevent data contamination by ensuring non-overlapping splits between training and test sets. Nonetheless, given the public availability of the articles online, there remains a possibility that some test data may indirectly overlap with training data due to external access and prior exposure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Ethical issues</head><p>This task is aimed at testing the factual knowledge which LLMs acquire during their training process, whose objective is language modelling. This task should not suggest, or stimulate, that LLMs should commonly be used as knowledge bases or as reliable sources of factual information. The investigation underlying this challenge is research-oriented, aimed at a better understanding of LLMs' abilities, and possibly suggest ways to discern when models might be providing more or less reliable knowledge and possibly making them more transparent in their generated output.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9.">Data license and copyright issues</head><p>Access to the data is granted for the evaluation but cannot be shared publicly at the moment, also for reasons related to data contamination. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Examples of Good titles selected by professional journalists</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C. Composition of the datasets used to train the classifiers</head><p>The dataset we used as a source of material for both the NS and HA classifiers is taken from "Testimole" <ref type="bibr" target="#b15">[16]</ref>, a massive collection of Italian web scraping data that includes a "blogs" subset containing, as of November 2024, more than 2.8 million posts from various online blogs and websites. From the original 2.8 million rows, we obtained a much smaller dataset by filtering articles coming from sources that are, to our judgement, more similar to professional media outlets. After this selection process, which yielded a total of 715,335 articles, we filtered out articles written in languages different than Italian by using the "FastText Lang ID" field already present in Testimole. After the foreign-languages pruning the count of articles was 293,518 articles. Finally, we discarded all the rows whose article was shorter than 350 characters to arrive to a final dataset size of 264,455 articles. In the following section, this dataset will be referred as "testimole-subset". In order to increase the diversity of data for the HA Classifier, we added to this dataset a collection of 432.000 articles taken from the professional Italian media outlet "Il Fatto Quotidiano": we had to add this source manually because the articles were missing from the original Testimole dataset due to a scraping issue. In the section of HA Classifier, we will refer to this additional subset as "testimole-subset-auxiliary". Finally, we are going to refer to the small subset of Galileo used in the testing process as "experimental-dataset". The experimental dataset contains 3007 original headlines from "Galileo" and 3007 headlines generated using Phi 3.5 Mini Instruct from the same subset of Galileo's articles.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>D. NS Classifier</head><p>For the NS Classifier, we decided to split the testimolesubset dataset in two sets: 60% of the dataset was kept with the original headline ("natural") while in the remaining 40% the original headline was substituted with a generated one ("synthetic"). The original headline is kept as a reference as a separate column in the dataset. Specifically, we generated 93,921 headlines and kept 132,227 original headlines. There is no contamination between generated and original headlines: no synthetic headlines were generated for headlines that are present in the dataset with the "natural" label. The dataset was then divided in "test" (45230 entries, x natural, x syntethic) and "train" (180918 entries, 105885 natural, 75033 synthetic) split for training.</p><p>For the generation, we ran Ollama on different models using the same prompt adopted for the evaluation. In Table <ref type="table" target="#tab_3">2</ref> you can see the amount of generated headlines for each model used.</p><p>The classifier was created using Hugging Face's transformers library. We initialized the model using AutoModelForSequenceClassification and trained the model using a binary cross-entropy loss function (BCEWithLogitsLoss).</p><p>Training was conducted with a batch size of 32, a learning rate of 2 × 10 -5}, and a warmup ratio of 0.1 to help stabilize early training. A linear learning rate scheduler and the $AdamW$ optimizer with gradient clipping were employed to manage learning stability. We also implemented early stopping, monitoring the F1 score to save the best model checkpoint and halt training if the model failed to improve over multiple epochs. The resulting model obtained a 95% of accuracy on the test set. Accuracy is measured as the number of correctly guessed labels divided for the total number of examples. The threshold to decide for a positive or negative label was set at 0.5. Using a continuos score instead of the threshold led to the same result, for this reason we decided to kept only accuracy in this report.</p><p>After having tested the model, we decided to further train it on the test set in order to have an improved model to be used for the CALAMITA task.</p><p>We then tested this further trained model on the smaller "experimental-dataset" dataset containing 3007 natural and 3007 synthetic headlines coming from the Galileo dataset. This evaluation obtained an accuracy of 87%</p><p>While initially we directly used PyTorch to train the experimental versions of the model, we then decided for simplicity to adopt the HuggingFace transformer library to easily upload the model on the HuggingFace hub. The further trained version of model is available at the address: https://huggingface.co/mrinaldi/flash-it-nsclassifier-fpt</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>E. HA Classifier</head><p>In order to build the HA Classifier we first computed, for each article contained in the "testimole-subset" dataset, the embedding of the article's text using SentenceBert with an Italian model <ref type="foot" target="#foot_4">6</ref> and added the embedding to a new column in the dataset. Then, we paired each article (source) of the dataset with the article (target) having the highest cosine similarity between the embeddings. After the pairing, both source and target were marked as "used" so that each article can appear no more than one time in the resulting dataset, either as a source or as a target. The resulting dataset<ref type="foot" target="#foot_5">7</ref> has 6 columns:</p><p>• Anchor: the body of the "source" article • Negative: the original title of the "target" article • Cosine similarity: the Cosine Similarity between the source's and target's embeddings computed on their texts • Url positive: the URL of the source article, it can be used as a key to find the original article in the Testimole dataset • Url negative: the URL of the target article Given the procedure employed for generating this dataset, the resulting number of row is halved so that, starting from the original 256530 entries in the "testimole-subset" dataset we obtained 128265 entries, divided into 102600 train entries and 25665 test entries. We believe that using the cosine similarity instead of randomly shuffling the articles can improve the performance of the classifier by increasing the difficulty of the task. Results with a classifier trained on randomly paired articles is present in the table below.</p><p>The classifier was created using Sentence-BERT, specifically by initializing the model with the SentenceTransformer class from the sentence_transformers library, using a pre-trained Italian model<ref type="foot" target="#foot_6">8</ref> . To fine-tune this model, we employed a TripletLoss function to enhance similarity-based ranking in embedding space. The triplet loss was the optimal choice given our dataset because it requires an anchor, a positive and a negative example. The goal of the triplet loss is to maximize the distance between the anchor and the negative example while at the same time minimize the distance between the anchor and the positive example. In this way, we encouraged the formation of meaningful embeddings tailored to minimize the distance between an article and a title coherent with its content, notwithstanding the 512 token length limitation.</p><p>Training was conducted over three epochs with a batch size of 64 for training and 16 for evaluation, using a learning rate of 2 × 10 -5} and a warmup ratio of 0.1 to stabilize initial training steps. We used the $SentenceTransformerTrainingArguments$ to configure training, applying half-precision floatingpoint (fp16) to speed up processing. An evaluation was performed every 1,000 steps to monitor model performance, with checkpoints saved periodically to retain the best-performing model. We kept the "margin" value at "5" following the documentation of SentenceBert. <ref type="foot" target="#foot_7">9</ref>The resulting classifier outputs a score representing the alignment between the article and its headline.</p><p>After having trained the HA Classifier on the "testimole-subset" dataset, we decided to use an additional dataset (testimole-auxilliary) to further improve the classifier. Testimole-Auxiliary, halved due to matching, has 216562 articles of which 108281 were used as train and 108281 as test. The same procedure used for testimole-subset was applied to testimole-auxilliary. In the following page we present a table summing up the results of the various models on the test datasets.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>•</head><label></label><figDesc>Title: the title of the article • Source: the name of the website • Date: the publishing date of the article • Author: the author of the article, if present • URL: the Internet address of the article • Text: the body of the article • ID: a unique identifier of the article</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Distribution of articles by token count in the Galileo subset.</figDesc><graphic coords="3,302.62,285.61,203.36,152.52" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Distribution of articles by token count in the Ansa subset.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>B. Examples of Model Outputs in Preliminary Evaluation</head><label></label><figDesc></figDesc><table><row><cell>Original Title (Italian)</cell><cell>LLM Generated Title (Italian)</cell></row><row><cell cols="2">No allo studente flessibile • Nella Via Lattea c'è un oggetto misterioso, è ve-Riforma Universitaria: Studenti in Rivolta contro la Selettività e la Flessibilità No to the flexible student University reform: Students in revolt against selectivity and flexibility locissimo Alzheimer, la proteina della speranza Nicastrina: nuova chiave per Alzheimer, scoperta in Calabria</cell></row><row><cell cols="2">• Nasce il gemello digitale del rischio ambientale Alzheimer, the protein of hope Nicastrin: new key for Alzheimer's, discovered in Calabria</cell></row><row><cell>in Italia Un laser estremo</cell><cell>Euv: Nuova Frontiera nella Manipolazione Molecolare con Luce Ultra-Sottile</cell></row><row><cell cols="2">• I cinque modi in cui il cervello invecchia An extreme laser EUV: New frontier in molecular manipulation with ultra-thin light</cell></row><row><cell cols="2">E' italiano il primo modello genetico • Covid-19, il mistero degli over 90 The first genetic model is Italian • A 44 e a 60 anni i due gradini chiave Creazione topi per studiare il ritardo mentale: scoperta genetica rivoluzionaria Creation of mice to study mental disability: revolutionary genetic discovery Carta d'identità per le balene Genetica della balena: Identificazione di specie protette nel commercio illegale dell'invecchiamento Identity card for whales Whale genetics: Identification of protected species in illegal trade • Palestra o snack? la scelta dipende da un messag-Scampati allo shock anafilattico Scoperto il fattore eNOS nella reazione anafilattica: speranza per una cura gero chimico Escaped from anaphylactic shock eNOS factor discovered in anaphylactic reaction: hope for a cure</cell></row><row><cell cols="2">• Dagli stadi alle spiagge, sono i salti a sincronizzare La casa in un clic Domotica: Software unificato per controllare tutti gli elettrodomestici, inno-</cell></row><row><cell>il ballo</cell><cell>vazione Cnr</cell></row><row><cell cols="2">The house in a click • Dalle rose alle melanzane, ecco i geni delle spine Home automation: Unified software to control all appliances, Cnr innovation Deforestazione senza fine Deforestazione in aumento in Amazzonia: il Brasile reagisce • Così il Covid accelera l'invecchiamento Endless deforestation Deforestation increasing in the Amazon: Brazil reacts • Uno zucchero naturale contro la calvizie, bene i "Più impegno contro il fumo" Bioetica del tabagismo: raccomandazioni per campagne di prevenzione e test sui topi interventi regolativi</cell></row><row><cell cols="2">• Scoperto nel cervello il circuito dell'effetto More commitment against smoking Bioethics of smoking: recommendations for prevention campaigns and regulatory</cell></row><row><cell>placebo</cell><cell>interventions</cell></row><row><cell cols="2">Una carovana contro l'Oleoducto • Pronto il Google Earth del cuore umano Protesta contro l'Oleodotto Crudos Pasos: Ecuador, ambientalismo e scontro politico • Una molecola può ringiovanire il sistema immu-A caravan against the pipeline Protest against the Crudos Pasos pipeline: Ecuador, environmentalism and polit-nitario ical clash • Scoperto il dizionario dei sinonimi e contrari del Analisi dell'egemonia americana Egemonia Americana: Stati Fuorilegge e Diritto Internazionale cervello Analysis of American hegemony American hegemony: Rogue states and international law</cell></row><row><cell cols="2">• Le farfalle nello stomaco non sono solo un modo Censura sui dati scientifici Scienza e Sicurezza: Accademia e Governo Discutono su Divulgazione nei</cell></row><row><cell>di dire</cell><cell>Lavori Scientifici</cell></row><row><cell cols="2">Censorship on scientific data • Pronto il primo orologio nucleare, il più preciso Science and security: Academia and government discuss disclosure in scientific works del mondo Siamo tutti africani: ricostruito il Dna degli DNA dell'Africa: Prove definitive dell'origine umana • Gli uccelli in volo si comportano come gli atomi antenati • L'Italia ritenta la sfida impossibile della geometria We are all Africans: reconstructed the DNA DNA of Africa: Definitive proof of human origin</cell></row><row><cell cols="2">• Le auto nel traffico come i batteri in cerca di cibo of ancestors</cell></row><row><cell cols="2">• Robot come alleati, trovata la chiave per collabo-Rubbia racconta Galileo divulgatore di Galileo Divulgatore: Scienza e Civiltà dal XVII al Mondo Moderno</cell></row><row><cell>rare con gli umani scienza</cell><cell></cell></row><row><cell cols="2">Rubbia tells Galileo as a science communica-• Dalle spugne di vetro grattacieli più sottili e re-Galileo communicator: Science and civilization from the 17th century to the tor modern world sistenti Galileo Galilei nel pensiero di Thomas Khun Kuhn e Galileo: Rivoluzione scientifica attraverso la meccanica e l'esperimento • L'IA non è razionale, fa ragionamenti non logici Galileo Galilei in the thought of Thomas Kuhn and Galileo: Scientific revolution through mechanics and experiment</cell></row><row><cell>Khun</cell><cell></cell></row><row><cell>Nella maglie della lingua</cell><cell>Realtà e Reti: L'Illusione di una Lingua Universale nella Scienza Cognitiva</cell></row><row><cell>In the mesh of the language</cell><cell>Reality and networks: The illusion of a universal language in cognitive science</cell></row><row><cell>Europa-America, la migrazione dei virus</cell><cell>Scambi e mutua scoperta: Storia sanitaria dell'unificazione globale</cell></row><row><cell>Europe-America, the migration of viruses</cell><cell>Exchanges and mutual discovery: Health history of global unification</cell></row><row><cell>Infiniti universi allacciati</cell><cell>Inflazione cosmica: un universo di mondi nascosti</cell></row><row><cell>Infinite universes connected</cell><cell>Cosmic inflation: a universe of hidden worlds</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Comparison of Original and LLM Generated Titles with Literal Translations.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>•</head><label></label><figDesc>Positive: the original title of the "source" article</figDesc><table><row><cell>Model</cell><cell>Count</cell><cell>Percentage</cell></row><row><cell>lama3.2:3b-instruct-fp16</cell><cell>51886</cell><cell>55.24%</cell></row><row><cell>qwen2.5:7b-instruct-q8_0</cell><cell>18418</cell><cell>19.61%</cell></row><row><cell>aya:8b-23-q8_0</cell><cell>17043</cell><cell>18.15%</cell></row><row><cell>mistral:7b-instruct-v0.3-q6_K</cell><cell>6312</cell><cell>6.72%</cell></row><row><cell>phi3.5:3.8b-mini-instruct-fp16</cell><cell>262</cell><cell>0.28%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Distribution of generated headlines by model</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.anthropic.com/news/claude-3-5-sonnet</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://openai.com/index/hello-gpt-4o/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://huggingface.co/datasets/mrinaldi/TestiMole</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://huggingface.co/nickprock/ sentence-bert-base-italian-xxl-uncased</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">https://huggingface.co/nickprock/ sentence-bert-base-italian-xxl-uncased</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">https://huggingface.co/datasets/mrinaldi/ flash-it-ha-dataset-cossim</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6">https://huggingface.co/nickprock/ sentence-bert-base-italian-xxl-uncased</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_7">https://sbert.net/docs/package_reference/sentence_transformer/ losses.html#tripletloss</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The authors would like to thank ANSA Scienza and Galileo, giornale di scienza -http:\www.galileonet.it for their interest in the GATTINA CALAMITA challenge and for the extremely valuable exchange of ideas that allowed us to shape a task of high potential impact in the field of journalism.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>(V. Patti) https://github.com/rosakun (M. Francis); https://github.com/mrinaldi97 (M. Rinaldi); https://github.com/Jj-source (J. Gili); https://github.com/malvinanissim (M. Nissim); https://github.com/vivpatti (V. Patti) 0009-0007-7638-9963 (M. Francis); 0009-0004-7488-8855 (M. Rinaldi); 0009-0007-1343-3760 (J. Gili); 0000-0001-5289-0971 (M. Nissim); 0000-0001-5991-370X (V. Patti)</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">CALAMITA: Challenge the Abilities of LAnguage Models in ITAlian</title>
		<author>
			<persName><forename type="first">G</forename><surname>Attanasio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Borazio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Francis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gili</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Musacchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nissim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rinaldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Scalena</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting>the 10th Italian Conference on Computational Linguistics (CLiC-it 2024)<address><addrLine>Pisa, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024-12-06">December 4 -December 6, 2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Suitable doesn&apos;t mean attractive. human-based evaluation of automatically generated headlines</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cafagna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">D</forename><surname>Mattei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bacciu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nissim</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-2481/paper13.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Sixth Italian Conference on Computational Linguistics</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">R</forename><surname>Bernardi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Navigli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</editor>
		<meeting>the Sixth Italian Conference on Computational Linguistics<address><addrLine>Bari, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">November 13-15, 2019. 2019</date>
			<biblScope unit="volume">2481</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Invisible to people but not to machines: Evaluation of style-aware headlinegeneration in absence of reliable human judgment</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">De</forename><surname>Mattei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cafagna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Dell'orletta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nissim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twelfth Language Resources and Evaluation Conference</title>
				<meeting>the Twelfth Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="6709" to="6717" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Pens: A dataset and generic framework for personalized news headline generation</title>
		<author>
			<persName><forename type="first">X</forename><surname>Ao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Qiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing</title>
		<title level="s">Long Papers</title>
		<meeting>the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="82" to="92" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Duan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Shou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cao</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2004.01401</idno>
		<title level="m">Xglue: A new benchmark dataset for cross-lingual pretraining, understanding and generation</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Harnessing the power of LLMs: Evaluating human-AI text co-creation through the lens of news headline generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smith-Renner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tetreault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jaimes</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.findings-emnlp.217</idno>
		<ptr target="https://aclanthology.org/2023.findings-emnlp.217.doi:10.18653/v1/2023.findings-emnlp.217" />
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2023, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">H</forename><surname>Bouamor</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Pino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Bali</surname></persName>
		</editor>
		<meeting><address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="3321" to="3339" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">A neural attention model for abstractive sentence summarization</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rush</surname></persName>
		</author>
		<idno>CoRR, abs/1509.00685</idno>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv Preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">ROUGE: A package for automatic evaluation of summaries</title>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W04-1013" />
	</analytic>
	<monogr>
		<title level="m">Text Summarization Branches Out, Association for Computational Linguistics</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="74" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1908.10084" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Generating user-engaging news headlines</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 61st Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="3265" to="3280" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Bleu: a method for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Papineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roukos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-J</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th annual meeting of the Association for Computational Linguistics</title>
				<meeting>the 40th annual meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="311" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The meteor metric for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lavie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Denkowski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine translation</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="105" to="115" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Rei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stewart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Farinha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lavie</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2009.09025</idno>
		<title level="m">Comet: A neural framework for mt evaluation</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Towards unified uni-and multi-modal news headline generation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Krubiński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pecina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EACL 2024</title>
				<imprint>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="437" to="450" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">M</forename><surname>Rinaldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Testimole</forename></persName>
		</author>
		<ptr target="https://huggingface.co/datasets/mrinaldi/TestiMole" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
