<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>at TA1C 2025: Leveraging Large Language Models for Identifying and Spoiling Clickbait in Spanish Language</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tomás Bernal-Beltrán</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ronghao Pan</string-name>
          <email>ronghao.pan@um.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Antonio García-Díaz</string-name>
          <email>joseantonio.garcia8@um.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rafael Valencia-García</string-name>
          <email>valencia@um.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Clickbait Identification, Clickbait spoiling, In-context Learning, Prompt-Tuning, Natural Language Processing</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Facultad de Informática, Universidad de Murcia, Campus de Espinardo</institution>
          ,
          <addr-line>30100</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Clickbait headlines are designed to attract attention by creating an information gap, often prioritizing engagement over transparency. Their widespread use on social media platforms has raised concerns about their impact on the credibility of digital journalism and the spread of misinformation. This paper describes our participation in the TA1C 2025 shared task on clickbait detection and spoiling in Spanish. For the detection subtask, we developed an ensemble-based approach that combines three fine-tuned encoder-only Transformer models (MarIA, BERTIN and ALBETO) with the decoder-only Large Language Model Gemma-2-2B-it. We fine-tuned this last model using QLoRA for eficient adaptation. In this subtask, our system achieved the highest score of all the participants, demonstrating the efectiveness of combining multiple architectures under a unified classification framework. For the spoiling subtask, we proposed a two-step, zero-shot pipeline based entirely on in-context learning with Gemma-2-2B-it. In the first stage, the model generates a guiding question from the headline. In the second stage, it generates a concise spoiler by answering the question using the article body. While our system did not achieve the best performance in this subtask, it highlights the potential of prompt-only approaches for information gap resolution without any task-specific training. Our results demonstrate that, while zero-shot prompting can deliver competitive performance with minimal resources, combining it with supervised signals or hybrid techniques may be essential for more complex generation tasks, such as clickbait spoiling.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Social media platforms; such as Facebook, X, TikTok, YouTube and Instagram, have become the main
source of information for much of the population. They provide instant access to news, opinions
and social interactions, and facilitate the proliferation of clickbait headlines. These are sensationalist
headlines designed to exploit readers’ curiosity and encourage them to click on a link. However,
clickbait can also manifest in other ways, such as misleading thumbnails or vague summaries. This
has significantly transformed the way users interact with news on digital platforms. Designed to
capture attention and encourage clicks, these headlines often sacrifice accuracy and informational
transparency, which can undermine public trust in the media [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. While certain clickbait features can
increase user engagement on platforms such as Facebook and X, studies have also found that this type
of content decreases the perceived credibility of news websites [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Moreover, repeated exposure to
misleading headlines can contribute to the spread of misinformation, as users tend to share content
without verifying its accuracy.
      </p>
      <p>Therefore, there is a critical need for efective mechanisms to identify and spoil this type of news,
in order to mitigate its impact on the perceived credibility of news websites. However, due to the
huge volume of content produced every second on social networks and news websites, performing this</p>
      <p>CEUR
Workshop</p>
      <p>ISSN1613-0073
identification manually becomes increasingly impractical. These processes are not only time-consuming
but also resource-intensive, which underscores the need for automated and scalable solutions.</p>
      <p>
        To address these challenges, the research community has been developing automated clickbait
detection systems that combine advanced Natural Language Processing (NLP) with machine learning
(ML) and deep learning (DL) techniques. Traditional clickbait detection methods analyze the linguistic
features of headlines in an attempt to identify common clickbait patterns. For example, DL models
with sentence embeddings have been employed to detect clickbait in various languages, including
lowresource languages such as Urdu [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and Indonesian [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Additionally, approaches combining semantic
analysis and ML techniques have been explored to enhance the accuracy of identifying misleading
headlines [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. These advancements are essential for curbing the spread of clickbait, mitigating the
potential misinformation it can cause, and restoring people trust in digital media.
      </p>
      <p>
        The TA1C (Te Ahorré Un Click) Clickbait detection and spoiling in Spanish [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] shared task; which
is part of IberLEF 2025 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]), aims to eficiently identify and generate short texts (spoils) of Spanish
language news articles that have previously been identified as clickbait. The overarching goal is to
reduce redundancy in clickbait impact mitigation eforts by detecting which news articles need to
be summarized/spoiled, and by satisfying users’ curiosity and filling the information gap created by
headlines. The task is divided into two “subtasks”: (1) Clickbait detection. Given a tweet that links
to a news article, the goal is to determine whether the article’s content is clickbait; and (2) Clickbait
spoiling. Given a clickbait teaser (tweet and title) and the corresponding news article, the goal is to
generate or extract from the article, a short text that fills the information gap as concisely as possible
(280 characters maximum) to satisfy the curiosity generated, or otherwise indicate that the article does
not address it.
      </p>
      <p>
        We participated in both subtasks of this shared task. For the Clickbait detection subtask, we propose
an approach based on an ensemble of four fine-tuned models, most of which are encoder-only models.
For the Clickbait spoiling subtask, we propose an approach based on the in-context learning capabilities
of Large Language Models (LLMs). Specifically, we use the Gemma-2-2B-it [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] model with zero-shot
learning techniques. We formulated the task as a Question Answering (QA) problem, for which the
relevant questions had to be generated in advance. To this end, we used zero-shot prompting techniques
with the Gemma-2-2B-it model, providing it with the news headline along with specific instructions to
generate the corresponding question. Then, we addressed the generation of the spoiler by once again
applying zero-shot prompting techniques using the same model. In this phase, the model was provided
with the subtitle and body of the news article, along with detailed instructions, to generate the text
revealing the information hidden behind the headline.
      </p>
      <p>The rest of this paper is organized as follows: Section 2 reviews related work and recent advances
relevant to the clickbait detection and spoiling task. Section 3 presents our system architecture and
describes the approach used for both subtasks, as well as the prompting strategies employed. Section 4
outlines the dataset, preprocessing steps and evaluation metrics. Section 5 reports and discusses the
results obtained. Section 6 summarizes our findings and outlines possible directions for future research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background Information</title>
      <p>The increasing volume of clickbait content generated on social networks and news websites, coupled
with the critical need for efective mechanisms to identify and spoil this type of news, has driven the
development of the automation of the clickbait detection and spoiling process. Two key steps in this
process are the detection of whether a news article is designed to exploit readers’ curiosity and drive
clicks; that is, if it’s clickbait, (Clickbait detection) and the generation or extraction of a short text that fills
the information gap, satisfying the curiosity generated by the misleading headline (Clickbait spoiling).
These tasks have been extensively studied in recent literature in both in monolingual and multilingual
contexts with the aim of creating robust systems that can operate eficiently across languages and
domains, including informal domains such as social networks.</p>
      <p>
        Early approaches tackled clickbait detection as a binary classification problem, whereby a headline,
and sometimes its accompanying content, was analysed to determine whether or not it was clickbait.
Initial eforts involved curating datasets by targeting publishers known for their clickbait content,
such as BuzzFeed, Hufington Post and Upworthy, or by using crowdsourcing for data annotation. For
example, in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] the authors compiled 15.000 news headlines labeled via a publisher-based heuristic and
in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], the authors used article informality cues to detect clickbait. However, these early datasets were
biased, for example, because they relied on the reputation of publishers, and many publishers with a
low reputation still publish significant amounts of news without resorting to clickbait teaser messages.
      </p>
      <p>
        In recent years, deep learning techniques have achieved state-of-the-art results in clickbait detection.
Researchers have experimented with a variety of approaches. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the authors use a hybrid
categorization approach that integrates diferent features, sentence structure and clustering. In [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], the
authors propose a detection model that integrates headline semantic and POS tag information.
      </p>
      <p>
        An important and unique aspect of clickbait detection is headline–article consistency. Clickbait often
occurs when a headline does not match the content, or when key information is withheld. To address
this issue, summarization-assisted methods have been proposed. For example, in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] the authors adopt
this approach by employing prompt-based tuning. They generate a summary of the article using a
pre-trained language model and then input the headline and the summary into a prompt that instructs
the model to evaluate whether the headline is clickbait.
      </p>
      <p>
        Another recent trend is the use of ensemble models. Given the variety of clickbait styles, ensembles
of classifiers can improve robustness. For example, combining encoder-only transformers (for headline
classification) with models that consider auxiliary information, such as the tweet text or metadata,
can boost accuracy. In [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], the authors proposed a hybrid approach that integrates textual content
with auxiliary features such as sentiment scores and engagement metrics. Their ensemble model,
which combines transformer-based language models with stacking classifiers, demonstrated enhanced
performance in identifying misinformation on Twitter.
      </p>
      <p>
        The task of automatic clickbait spoiling was first introduced in [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], where the authors compiled the
Webis Clickbait spoiling Corpus 2022. This corpus consists of 5.000 social media posts from Twitter,
Reddit and Facebook, each of which is paired with a manually crafted spoiler text that reveals the
information behind the tease. The proposed pipeline treats the spoiling task as a two-step process: (1)
classifying the type of spoiler required. Either a short phrase (extracted directly from the text) or a
longer passage (extracted from diferent parts of the text); and (2) generating the spoiler text itself. For
the spoiler generation step, the authors employed a QA approach, reformulating the clickbait article
headline as a question and using an extractive QA model to retrieve the answer from the article body.
      </p>
      <p>
        Subsequently, in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], the authors introduced the Jack‑Ryder system in SemEval 2023 Task 5 [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ],
which embraced a zero-shot QA approach. Their method automatically rewrites clickbait headlines as
questions and uses pre-trained QA models to extract spoilers directly from the text, without requiring
any additional training. Furthermore, they strategically reorder the sentences within each article based
on semantic similarity and select the most suitable model for each spoiler type, achieving competitive
results purely through prompting and selection strategies.
      </p>
      <p>
        An alternative strategy to solve this task is to treat it as a sequence-to-sequence learning problem,
similar to summarization, in which LLMs are fine tuned to map the article body (or part of it) onto the
spoiler text. For instance, [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] adopted this approach in the same SemEval Task, experimenting with
various T5 models and finding that a T5-Large model fine-tuned for article-to-spoiler generation task
could efectively generate spoilers when provided with the article content as input. Another example
is [19], where the authors adopted a zero-shot prompt setup to evaluate a diverse set of LLMs in the
clickbait summarization task. These studies have shown that abstractive approaches, such as QA or
summarization, can be efective, particularly when an initial step is taken to determine the expected
answer format.
      </p>
      <p>The emergence of LLMs has recently opened up new possibilities for addressing the spoiling task
through their in-context learning capabilities. LLMs can perform tasks without explicit fine-tuning,
relying instead on carefully designed prompts with examples provided at inference time. This approach
is particularly well-suited to generation tasks such as clickbait spoiling, where the model is asked to
synthesize concise and informative answers from long article texts.</p>
      <p>
        In fact, researchers have started using LLMs by simply instructing them to produce spoilers to
clickbait articles. The QA system described earlier in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] is an example of this approach: the authors
used a text generation model as part of a zero-shot prompting spoiler extraction pipeline, crafting
prompts such as “Provide the answer to the headline: [headline]?” to guide the model. Similarly, in
[20], the authors used a prompt-based approach with a pre-trained language model, providing it with
the article body alongside a directive to answer the headline’s question; essentially building a prompt
that instructs the model to perform the task like a user of “I saved you a click” website would do:
reading the article and extracting the key detail. These studies demonstrated that LLMs can perform the
spoiler extraction task efectively when given the right context and instructions, even without explicit
training on a clickbait dataset or the clickbait spoiling task. This approach achieves performance that is
competitive with fine-tuned systems while requiring significantly lower computational resources.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. System Overview</title>
      <p>
        The ensemble comprises four Transformer-based models: three encoder-only models: MarIA [21],
BERTIN [22] and ALBETO [23], and one decoder-only large language model: Gemma-2-2B-it [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Each
model was fine-tuned using specific hyperparameters, including the number of epochs, the batch size,
the learning rate and the weight decay. These values were individually selected for each model via grid
search within the following ranges: epochs ∈ [
        <xref ref-type="bibr" rid="ref1 ref10 ref15 ref5">1, 5, 10, 15, 20</xref>
        ], batch size ∈ [
        <xref ref-type="bibr" rid="ref1 ref16 ref2 ref4 ref8">1, 2, 4, 8, 16, 32</xref>
        ], learning
rate ∈ [1e-5, 1e-6, 3e-5, 5e-5] and weight decay ∈ [0, 0.01, 0.05, 0.1].
      </p>
      <p>We used the HuggingFace Transformers library to perform the fine-tuning for the detection subtask,
formulating it as a binary text classification problem. Each piece of input text was assigned a single
label: 1 for clickbait news articles, and 0 for non clickbait ones. In the case of encoder-only models,
the classification head added to the encoder consists of a single linear layer on top of the [CLS] token,
which produces logits for the two classes. For the decoder-only model, Gemma-2-2B-it, we adopted
a prompt-based classification approach. During fine-tuning, the model was trained to generate a
predefined label token in response to a natural language prompt containing the input headline. This
token is then mapped to the corresponding binary label.</p>
      <p>Fine-tuning was performed using the Trainer API, with cross-entropy loss as the optimization
objective. Standard preprocessing steps included tokenizing the input texts, truncating or padding them
to a fixed maximum length and batching them eficiently for GPU training. The tokenizer automatically
handled special tokens and padding, and no additional label masking was required since classification
was performed at the sequence level. For encoder-only models, fine-tuning was performed with full
precision. For Gemma-2-2B-it, we applied parameter-eficient fine-tuning using QLoRA [ 24], which
allowed us to adapt the model to the classification task while significantly reducing memory usage and
training time.</p>
      <p>
        Figure 2 illustrates the overall system architecture of the spoiling subtask. Our solution follows a
two-step pipeline, entirely based on zero-shot prompting using the Gemma-2-2B-it model [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], without
any additional fine-tuning.
      </p>
      <p>Zero-shot prompt</p>
      <p>Teaser text (tweet) Article headline
Gamestop: cómo inversores aficionados de Gamestop: cómo inversores aficionados de
ganaRreodnd[i.t..s]eesetnofr"ennotaersonunaaWmaoldSatrpeaestayjera" Reddigtasneaeronnfre(pnotarreolnmaoWmaelntSot)reet y</p>
      <p>Tokenizer
Zero-shot prompt
Generated Question</p>
      <p>Article body
Wal Street vive una batala inusual. Los
protagonistas son, por un lado [...] sus</p>
      <p>decisiones especulativas.</p>
      <p>Tokenizer
E0
...</p>
      <p>Embeddings</p>
      <p>En</p>
      <p>Output
2nd step</p>
      <p>Output
Clickbait Article Spoil</p>
      <p>In the first stage, since we formulated the spoiling subtask as a QA problem, our system addresses
the generation of the questions that will be used to instruct the model to generate the spoiler. We then
prompt the model with carefully crafted instructions designed to generate an explicit question that
captures the information gap created by the clickbait article headline.</p>
      <p>In the second stage, we use a diferent zero-shot prompt to generate the actual spoiler for the clickbait
article. The model receives the previously generated question, as well as the subtitle and body of the
news article, as input. The prompt instructs the model to generate a concise and informative response
that addresses the question and satisfies the reader’s curiosity. This two-step process enables the system
to separate the reformulation of the headline from the generation of the answer, resulting in spoilers
that are more focused and relevant.</p>
      <p>
        We used the HuggingFace Transformers library to implement our solution for the spoiling subtask.
Unlike the detection subtask, we did not perform any fine-tuning. Instead, we relied entirely on
zeroshot prompting with the Gemma-2-2B-it model [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In this process, the input text was pre-processed
by being lowercased and having extra whitespace trimmed, and then tokenized using the model’s
tokenizer.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <p>For the detection subtask, our experiments were based exclusively on the training and development
sets provided by the task organizers. Table 1 shows the number of examples and class distribution for
each split.</p>
      <p>To build the ensemble used to solve this subtask, we fine-tuned four diferent pre-trained language
models, each with a diferent set of hyperparameters. For MarIA, fine-tuning was performed for 10
epochs with a learning rate of 1e-5, no weight decay and a batch size of 16. BERTIN was fine-tuned
for 20 epochs with a learning rate of 1e-6, no weight decay and a batch size of 16. ALBETO was
ifne-tuned for 15 epochs with a learning rate of 3e-5, a weight decay of 0.01 and a batch size of 32.
For Gemma-2-2B-it, we applied parameter-eficient fine-tuning based on QLoRA, training the model
for 10 epochs with a learning rate of 5e-5, a weight decay of 0.05 and a batch size of 4. LoRA-specific
hyperparameters included a rank (r) of 8, an LoRA alpha of 16 and a dropout rate of 0.1.</p>
      <p>For the spoiling subtask, our experiments were based exclusively on the training and development
sets provided by the task organizers. Table 2 shows the number of examples for each split.</p>
      <p>The first step in the pipeline designed to solve the spoiling subtask is to generate the questions that
will guide the model during the spoiler generation stage. To achieve this, we used the prompt shown in
Figure 1, which includes the article headline and a set of instructions for the model. Text generation
was performed using the HuggingFace Transformers library’s pipeline utility. We set do_sample=False,
corresponding to greedy decoding. This means the model always selects the most probable token at
each generation step, ensuring the generated questions are coherent and consistent.</p>
      <sec id="sec-4-1">
        <title>Listing 1: Prompt used for question generation</title>
        <p>T a r e a : Dado un t í t u l o de n o t i c i a s e n s a c i o n a l i s t a en e s p a ñ o l , g e n e r a
una p r e g u n t a i n f o r m a t i v a y e s p e c í f i c a que un l e c t o r c u r i o s o p o d r í a
h a c e r s e t r a s l e e r l o . La p r e g u n t a d e b e s e r b r e v e , d i r e c t a y c a p t a r
l a e s e n c i a de l o que s e q u i e r e s a b e r .</p>
        <p>T í t u l o : { h e a d l i n e }</p>
        <p>The objective of the second step in this pipeline is to generate the actual spoiler for the clickbait
article. To achieve this, we used the prompt shown in Figure 2. This prompt includes the previously
generated question and the full article content (subtitle and body), as well as a set of instructions for the
model. Text generation was performed using the generate() method with stochastic decoding enabled
(do_sample=True). A combination of nucleus sampling (top_p=0.95), top-k sampling (top_k=10) and
temperature-based scaling (temperature=0.7 ) was used to enable diverse yet controlled responses. The
number of output tokens was limited to a maximum of 280 to match the task requirements. Decoding
was performed by removing the prompt and any special tokens to extract the final spoiler.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Listing 2: Prompt used for spoiler generation</title>
        <p>E r e s un e x p e r t o en d e t e c t a r y r e v e l a r e l c o n t e n i d o r e a l d e t r á s de
a r t í c u l o s p e r i o d í s t i c o s s e n s a c i o n a l i s t a s en e s p a ñ o l ( c l i c k b a i t ) .
Tu t a r e a e s l e e r l a p r e g u n t a i n d i c a d a y e l t e x t o d e l a r t í c u l o
c o m p l e t o , p a r a l u e g o r e s p o n d e r con una s o l a f r a s e b r e v e a d i c h a
p r e g u n t a b a s á n d o t e en e l c o n t e n i d o d e l a r t í c u l o .
∗ ∗ I n s t r u c c i o n e s : ∗ ∗
− La r e s p u e s t a d e b e s e r c l a r a , d i r e c t a , n e u t r a l y c o n c i s a , s i n
i n t r o d u c i r i n f o r m a c i ó n que no e s t é en e l a r t í c u l o .
− Usa s o l o una f r a s e s i n l i s t a s , s u b t í t u l o s n i a d o r n o s .
− No r e p i t a s e l c o n t e n i d o de l a p r e g u n t a .
− La r e s p u e s t a d e b e t e n e r como máximo 2 8 0 c a r a c t e r e s
( i n c l u y e n d o e s p a c i o s ) .
− S i e l a r t í c u l o no c o n t i e n e una r e s p u e s t a c l a r a , r e s p o n d e con :
” No hay r e s p u e s t a ” .</p>
        <p>P r e g u n t a : { g e n e r a t e d _ q u e s t i o n }
T e x t o : { a r t i c l e _ b o d y }</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>In this section, we present and analyze the results obtained by our top-performing submission on
the oficial test set. Throughout the development phase, we conducted multiple experiments, testing
diferent fine-tuning configurations and alternative model combinations for the ensemble used in the
detection subtask, and several prompt variations for the spoiling subtask. The results discussed here
correspond to the best-performing setup submitted for each subtask.</p>
      <p>Table 3 shows the oficial results of the detection subtask, ordered by overall F1-score. Our system
(tomasbernal01) achieved the highest F1-score, with a value of 0.81564, outperforming the second and
third-ranked submissions by a small margin. In terms of individual metrics, our system achieved a recall
of 0.87425 and a precision of 0.76439. These results demonstrate the efectiveness of our ensemble-based
approach, which combines multiple fine-tuned encoder-only models and a decoder-only LLM under a
unified classification framework.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and Further Work</title>
      <p>In this work, we presented our system for the TA1C 2025 shared task, which addresses both the
detection and spoiling subtasks. For the detection subtask, our ensemble-based approach combining three
ifne-tuned encoder-only models and a decoder-only LLM, achieved the highest score of all participants.
Integrating Gemma-2-2B-it using parameter-eficient fine-tuning based on QLoRA allowed us to
effectively incorporate a decoder-only architecture into a classification ensemble at low computational
cost.</p>
      <p>For the spoiling subtask, we proposed a fully zero-shot pipeline based on in-context learning with
Gemma-2-2B-it. Our system framed the task as a two-step QA problem. First, it generated a guiding
question from the news headline using prompting. Second, it generated the spoiler by conditioning on
both the generated question and the article content. Despite lacking any task-specific fine-tuning or
supervised examples, our approach achieved competitive performance, demonstrating the viability of
prompt-only setups for resolving information gaps in Spanish-language news.</p>
      <p>Although our system performed well in the detection subtask, the results of the spoiling subtask
suggest that zero-shot prompting alone may not consistently generate high-quality spoils, particularly
when the answer is implicit or scattered throughout the article.</p>
      <p>Several directions remain open for future work, particularly for the spoiling subtask, where we
plan to explore few-shot prompting. In particular, integrating example selection strategies, such as
those proposed in recent work on Spanish hate speech detection using Gemma models [25], may
improve performance compared to random few-shot or zero-shot setups. Another promising approach
is to apply multi-task learning strategies, which have been successful in related fields such as hate
speech detection [26]. This could unify subtasks such as question generation, spoiler generation and
spoiler-type prediction under a single model. Furthermore, we recognize the potential of incorporating
emotional and rhetorical analysis. Since many clickbait headlines rely on emotional manipulation
to drive engagement, integrating emotion recognition techniques could enhance both detection and
spoiler generation. In this regard, recent eforts such as the Spanish MEACorpus 2023 [ 27] open the
door to enriching our models with emotional and prosodic features, which may help characterize the
persuasive intent and psychological impact of clickbait more precisely.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work is part of the research project LaTe4PoliticES (PID2022-138099OB-I00) funded by
MCIN/AEI/10.13039/501100011033 and the European Fund for Regional Development (ERDF)-a way to
make Europe. Mr. Tomás Bernal-Beltrán is supported by University of Murcia through the predoctoral
programme.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used DeepL for grammatical and spelling correction.
After using this tool, the authors reviewed and edited the content as needed and takes full responsibility
for the publication’s content.
seq2seq models, in: Proceedings of the 17th International Workshop on Semantic Evaluation
(SemEval-2023), 2023, pp. 1650–1655.
[19] I. García-Ferrero, B. Altuna, Noticia: A clickbait article summarization dataset in spanish,
Procesamiento del Lenguaje Natural 73 (2024) 191–207.
[20] O. Johnson, B. Lou, J. Zhong, A. Kurenkov, Saved you a click: Automatically answering clickbait
titles, arXiv preprint arXiv:2212.08196 (2022).
[21] A. G. Fandiño, J. A. Estapé, M. Pàmies, J. L. Palao, J. S. Ocampo, C. P. Carrino, C. A. Oller,
C. R. Penagos, A. G. Agirre, M. Villegas, Maria: Spanish language models, Procesamiento del
Lenguaje Natural 68 (2022). URL: https://upcommons.upc.edu/handle/2117/367156#.YyMTB4X9A-0.
mendeley. doi:10.26342/2022- 68- 3.
[22] J. D. la Rosa y Eduardo G. Ponferrada y Manu Romero y Paulo Villegas y Pablo González de
Prado Salas y María Grandury, Bertin: Eficient pre-training of a spanish language model using
perplexity sampling, Procesamiento del Lenguaje Natural 68 (2022) 13–23. URL: http://journal.
sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6403.
[23] J. Cañete, S. Donoso, F. Bravo-Marquez, A. Carvallo, V. Araujo, Albeto and distilbeto: Lightweight
spanish language models. arxiv, arXiv preprint arXiv:2204.09145 (2022).
[24] T. Dettmers, A. Pagnoni, A. Holtzman, L. Zettlemoyer, Qlora: Eficient finetuning of quantized
llms, Advances in neural information processing systems 36 (2023) 10088–10115.
[25] R. Pan, J. A. García-Díaz, R. Valencia-García, Optimizing few-shot learning through a consistent
retrieval extraction system for hate speech detection, Procesamiento del Lenguaje Natural 74
(2025) 241–252.
[26] R. Pan, J. A. Garcia-Diaz, R. Valencia-Garcia, Spanish mtlhatecorpus 2023: Multi-task learning
for hate speech detection to identify speech type, target, target group and intensity, Computer
Standards &amp; Interfaces 94 (2025) 103990.
[27] R. Pan, J. A. García-Díaz, M. Á. Rodríguez-García, R. Valencia-García, Spanish meacorpus 2023:
A multimodal speech–text corpus for emotion analysis in spanish from natural environments,
Computer Standards &amp; Interfaces 90 (2024) 103856.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>N.</given-names>
            <surname>Firdaus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jumroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aziz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sumartono</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Purwanti</surname>
          </string-name>
          ,
          <article-title>The influence of social media, misinformation, and digital communication strategies on public perception and trust</article-title>
          ,
          <source>The Journal of Academic Science</source>
          <volume>1</volume>
          (
          <year>2024</year>
          )
          <fpage>131</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>A.-K. Jung</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Stieglitz</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Kissmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Mirbabaie</surname>
          </string-name>
          , T. Kroll,
          <article-title>Click me…! the influence of clickbait on user engagement in social media and the role of digital nudging</article-title>
          ,
          <source>Plos one 17</source>
          (
          <year>2022</year>
          )
          <article-title>e0266743</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Muqadas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. U.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ramzan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Naz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Alsahfi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Daud</surname>
          </string-name>
          ,
          <article-title>Deep learning and sentence embeddings for detection of clickbait news from online content</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>15</volume>
          (
          <year>2025</year>
          )
          <fpage>13251</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N. P. I.</given-names>
            <surname>Maharani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Purwarianti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Aji</surname>
          </string-name>
          ,
          <article-title>Low-resource clickbait spoiling for indonesian via question answering</article-title>
          ,
          <source>in: 2023 10th International Conference on Advanced Informatics: Concept, Theory and Application (ICAICTA)</source>
          , IEEE,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bronakowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Al-Khassaweneh</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Al Bataineh,
          <article-title>Automatic detection of clickbait headlines using semantic analysis and machine learning techniques</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>13</volume>
          (
          <year>2023</year>
          )
          <fpage>2456</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Mordecki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chiruzzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Laguna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Prada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rosá</surname>
          </string-name>
          , I. Sastre, G. Moncecchi, Overview of TA1C at IberLEF 2025:
          <article-title>Detecting and Spoiling Clickbait in Spanish-Language News</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>75</volume>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          , pp. -.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Rivière</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pathak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. G.</given-names>
            <surname>Sessa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hardin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhupatiraju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hussenot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mesnard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shahriari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ferret</surname>
          </string-name>
          , et al.,
          <source>Gemma</source>
          <volume>2</volume>
          :
          <article-title>Improving open language models at a practical size</article-title>
          ,
          <source>CoRR</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Paranjape</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kakarla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ganguly</surname>
          </string-name>
          ,
          <article-title>Stop clickbait: Detecting and preventing clickbaits in online news media</article-title>
          , in: 2016 IEEE/ACM international conference
          <article-title>on advances in social networks analysis and mining (ASONAM)</article-title>
          , IEEE,
          <year>2016</year>
          , pp.
          <fpage>9</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Biyani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Tsioutsiouliklis</surname>
          </string-name>
          , J. Blackmer, ”
          <article-title>8 amazing secrets for getting more clicks”: detecting clickbaits in news streams using article informality</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>30</volume>
          ,
          <year>2016</year>
          , pp. -.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Pujahari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Sisodia</surname>
          </string-name>
          ,
          <article-title>Clickbait detection using multiple categorisation techniques</article-title>
          ,
          <source>Journal of information science 47</source>
          (
          <year>2021</year>
          )
          <fpage>118</fpage>
          -
          <lpage>128</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , X. Wu,
          <article-title>Wcd: A new chinese online social media dataset for clickbait analysis and detection</article-title>
          ,
          <source>in: 2021 7th IEEE International Conference on Network Intelligence</source>
          and
          <article-title>Digital Content (IC-NIDC)</article-title>
          , IEEE,
          <year>2021</year>
          , pp.
          <fpage>368</fpage>
          -
          <lpage>372</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Qiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <article-title>Prompt-tuning for clickbait detection via text summarization</article-title>
          ,
          <source>CoRR</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Dar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hashmy</surname>
          </string-name>
          ,
          <article-title>A multimodal ensemble machine learning approach to covid-19 misinformation detection in twitter</article-title>
          ,
          <source>in: ITM Web of Conferences</source>
          , volume
          <volume>54</volume>
          ,
          <string-name>
            <given-names>EDP</given-names>
            <surname>Sciences</surname>
          </string-name>
          ,
          <year>2023</year>
          , p.
          <fpage>01015</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jurk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Clickbait spoiling via question answering and passage retrieval, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>7025</fpage>
          -
          <lpage>7036</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Wangsadirdja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pfister</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kobs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hotho</surname>
          </string-name>
          ,
          <article-title>Jack-ryder at semeval-2023 task 5: Zero-shot clickbait spoiling by rephrasing titles as questions</article-title>
          ,
          <source>in: Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1090</fpage>
          -
          <lpage>1095</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gollub</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          , M. Potthast, SemEval
          <article-title>-2023 task 5: Clickbait spoiling</article-title>
          , in: A.
          <string-name>
            <surname>K. Ojha</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Doğruöz</surname>
            , G. Da San Martino, H. Tayyar Madabushi,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , E. Sartori (Eds.),
          <source>Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>2275</fpage>
          -
          <lpage>2286</lpage>
          . URL: https: //aclanthology.org/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .312/. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .
          <fpage>312</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>T.</given-names>
            <surname>Bilgis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. B.</given-names>
            <surname>Bozdag</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bethard</surname>
          </string-name>
          , Gallagher at semeval
          <article-title>-2023 task 5: Tackling clickbait with</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>