<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the CLEF-2024 CheckThat! Lab Task 1 on Check-Worthiness Estimation of Multigenre Content</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Maram Hasanain</string-name>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Reem Suwaileh</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sanne Weering</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chengkai Li</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tommaso Caselli</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wajdi Zaghouani</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alberto Barrón-Cedeño</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Preslav Nakov</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Firoj Alam</string-name>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DIT, Università di Bologna</institution>
          ,
          <addr-line>Forlì</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Hamad Bin Khalifa University</institution>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Mohamed bin Zayed University of Artificial Intelligence</institution>
          ,
          <addr-line>UAE</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Northwestern University in Qatar</institution>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Qatar Computing Research Institute</institution>
          ,
          <addr-line>HBKU</addr-line>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of Groningen</institution>
          ,
          <country country="NL">Netherlands</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>University of Texas at Arlington</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We present an overview of the CheckThat! Lab 2024 Task 1, part of CLEF 2024. Task 1 involves determining whether a text item is check-worthy, with a special emphasis on COVID-19, political news, and political debates and speeches. It is conducted in three languages: Arabic, Dutch, and English. Additionally, Spanish was ofered for extra training data during the development phase. A total of 75 teams registered, with 37 teams submitting 236 runs and 17 teams submitting system description papers. Out of these, 13, 15 and 26 teams participated for Arabic, Dutch and English, respectively. Among these teams, the use of transformer pre-trained language models (PLMs) was the most frequent. A few teams also employed Large Language Models (LLMs). We provide a description of the dataset, the task setup, including evaluation settings, and a brief overview of the participating systems. As is customary in the CheckThat! Lab, we release all the datasets as well as the evaluation scripts to the research community. This will enable further research on identifying relevant check-worthy content that can assist various stakeholders, such as fact-checkers, journalists, and policymakers.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Check-worthiness</kwd>
        <kwd>fact-checking</kwd>
        <kwd>multilinguality</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Check-worthiness is a crucial component of the fact-checking pipeline. It helps to alleviate the burden
on fact-checkers by reducing the need to verify every claim posted or shared across multiple online and
social media platforms, which contain diferent types of content and modalities. This content can include
news reports, citizen journalism, political debates, and posts from social media platforms. Identifying
and debunking misleading claims is crucial to prevent the spread of misinformation, enabling individuals
to make informed decisions where false information could lead to harmful consequences. For example,
in critical areas such as health, finance, natural disasters and public policy, making well-informed
decisions is especially important.</p>
      <p>
        The CheckThat! 2024 lab was held in the framework of CLEF 2024 [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ].1 Figure 1 shows the full
CheckThat! identification and verification pipeline, highlighting the six tasks targeted in this seventh
edition of the lab: Task 1 on check-worthiness estimation, Task 2 on subjectivity, Task 3 on persuasion
technique detection (this paper), Task 4 on detecting hero, villain, and victim from memes, Task 5 on
rumor verification using evidence from authorities, and Task 6 on robustness of credibility assessment
with adversarial examples.
      </p>
      <p>
        In this paper, we describe Task 1, which asks to detect whether a given text snippet from multigenre
content, in a form of a tweet or a sentence from a political debate or speech, is worth fact-checking.
Checkworthiness estimation simplifies and speeds up the process of fact-checking by prioritizing more
important claims to be verified. In order to make that decision, one would need to consider questions
such as “does it contain a verifiable factual claim?” or “is it harmful?”, before deciding on the final
check-worthiness label [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>We provided manually annotated data in three languages: Arabic, Dutch, and English. Additionally,
we included Spanish as an extra dataset. Among the various languages, English was the most popular
target for participants. Across the submitted systems, pre-trained language models (PLMs) were widely
used, with BERT, RoBERTa, and XLM-RoBERTa being the most popular models. Moreover, some teams
used large language models (LLMs). The top-ranked systems also employed data augmentation and
additional preprocessing steps.</p>
      <p>The remainder of the paper is organized as follows: Section 2 describes the datasets released with the
task. We present the evaluation setup in section 3. Section 4 discusses the system submissions and the
oficial results. Section 5 presents some related work. Finally, we provide some concluding remarks in
section 6.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Datasets</title>
      <p>
        The dataset contains multigenre content in Arabic, English, Dutch, and Spanish. The Spanish subset
was only ofered for training purposes. The evaluation focuses on the other three languages. For all
languages but English, the dataset consists of tweets collected using keywords related to a variety
of topics, such as COVID-19 and vaccines, climate change, political news and the war on Gaza. The
choice of topics was language-specific and was based on current events at diferent points of time
when the dataset was being constructed. Additionally, the Spanish subset included transcriptions from
Spanish politicians, and the subset was manually annotated by professional journalists who are experts
in fact-checking. To annotate Arabic and Dutch data, we followed the scheme described by Alam et al.
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. As for the English subset, it was sourced from the annotated dataset described by Arslan et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ],
and consists of transcribed sentences from candidates during the US Presidential election debates.
      </p>
      <p>
        We create the training, development and dev-test subsets for the 2024 edition by re-using all the
data released in 2023 (or 2022 when the language was not run in the 2023 edition). Regarding the
testing data, for Arabic we collected tweets using keywords relevant to the war on Gaza, that started
in October 2023. For Dutch, we collected 1 messages between January 2021 and December 2022 on
climate change and its associated debate. The English test set was constructed by manually annotating
transcribed sentences that did not appear in Arslan et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Table 1 shows statistics for all languages
and partitions.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Evaluation Settings</title>
      <p>We provided training, development, and dev-test subsets. The latter was intended to allow participants
to validate their systems internally, while they could use the development set for hyper-parameter
tuning and model selection. The test set was used for the final evaluation and ranking. The participants
were allowed to submit multiple runs on the test set (without seeing the scores), and the last valid run
was considered as oficial.</p>
      <p>This is a binary classification task and we evaluate it on the basis of the F 1-measure on the
checkworthiness class (yes) to account for class imbalance. The data and the evaluation scripts are available
online.2 The submission system was hosted on the CodaLab platform.3</p>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Overview of the Systems</title>
      <p>A total of 13, 15 and 26 teams submitted systems for Arabic, Dutch, and English, respectively. Table
3 reports the performance results for all systems and languages. For all languages, the participating
systems outperformed the baseline, except for one team in Arabic and two teams in Dutch.</p>
      <p>
        Table 2 summarizes the approaches. Transformers were most popular. Some teams used
languagespecific transformers, while others opted for multilingual ones. Several teams also used large language
models including variations of LLaMA, Mistral, Mixtral, and GPT. Standard preprocessing and data
augmentation were also very common. Below, we briefly describe the systems across all languages.
Team Fired_from_NLP [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] leveraged various model groups: Random Forest, SVM, and XGBoost;
deep learning models such as LSTM and Bi-LSTM; and pre-trained language models (PLMs) including
AraBERT for Arabic, RobBERT for Dutch, BERT-uncased for English, and Multilingual-BERT-uncased
for all three languages. They trained and fine-tuned the models using the original datasets. Experiments
showed that PLMs outperformed all other models.
      </p>
      <p>
        Team Fraunhofer SIT [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] proposed an adapter fusion approach that combines a task adapter model
with a Named Entity Recognition (NER) adapter, ofering a resource-eficient alternative to fully
finetuned PLMs. The task adapter was trained using the original training data without any preprocessing
or cleaning. This method demonstrated superior performance and achieved the third place in the task.
Team Mirela [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] used DistilBERT-multilingual and XLM-RoBERTa-base PLMs. DistilBERT-multilingual
was chosen for its lightweight and fast performance during inference, as well as its low computational
training requirements. XLM-RoBERTa-base was selected due to its pre-training on 100 languages,
achieving state-of-the-art performance in various NLP tasks in multilingual setups. Both models were
ifnetuned on the original training data for English, Spanish, Arabic, and Dutch.
      </p>
      <p>
        Team SSN-NLP [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] used a range of machine learning algorithms, including Support Vector Machine
(SVM), Random Forest Classifier, Logistic Regression, XGBoost Classifier, CatBoost Classifier, K-Nearest
Neighbors (KNN), and Passive Aggressive Classifier. Additionally, they fine-tuned several PLMs,
including BERT-base-uncased, RoBERTa-base, XLM-RoBERTa-base, and DeBERTa-v3-base. Hyperparameters
2https://gitlab.com/checkthat_lab/clef2024-checkthat-lab/-/tree/main/task1
3https://codalab.lisn.upsaclay.fr/competitions/18893
were optimized using GridSearchCV on the original data. Their preprocessing pipeline included text
cleaning, tokenization, stopword removal, punctuation removal, URL removal, and spelling correction.
For feature extraction, they used POS tagging and dependency parsing. These features were aggregated
into vectors and combined with sentence embeddings generated using the Sentence-BERT PLM. The
combined features were then normalized and reduced using Principal Component Analysis (PCA) to
minimize computational requirements.
      </p>
      <p>
        Team FactFinders [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] fine-tuned Llama2 7b on the original training data, using prompts generated
by ChatGPT. A similar performance was achieved through a 2-step data pruning technique, which
reduced the training data by 44% without compromising performance. The pruning involved filtering
informative sentences and applying the Condensed Nearest Neighbor undersampling technique. Despite
a slight performance drop (&lt;0.5%) with the pruned dataset, results were submitted using the model
ifne-tuned on the original data. The models showed variability in results across diferent runs, so the
ifnal results were based on the majority of five iterations. Other open-source LLMs, such as Mistral,
Mixtral, Llama2 13b, Llama3 8b, and CommandR, were also evaluated. Mixtral achieved the highest
F1-score in the dev-test phase, followed by Llama2 7b. Due to training time considerations, Llama2 7b
was used for the remainder of the study. Experiments with data expansion techniques yielded high
precision but lower recall models.
      </p>
      <p>
        Team SemanticCuetSync [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] fine-tuned language specific models such as RoBERTa, AraBERT,
DistilBERT for English, Arabic and Dutch, respectively.
      </p>
      <p>
        Team Checker Hacker [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] employed an ensemble approach integrating BERT-base-uncased and
XLMRoBERTa to improve the detection of check-worthy claims. Preprocessing steps, including tokenization
and normalization, were implemented, along with data augmentation techniques to ensure the model
was exposed to varied textual representations.
      </p>
      <p>
        Team IAI Group [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] trained several PLMs. For English, RoBERTa-Large was fine-tuned, and for Dutch
and Arabic, XLM-RoBERTa and GPT-3.5-Turbo were fine-tuned. The best models among them were
selected based on their performance on the dev-test subsets. They reported that in some cases, GPT-4
in a zero-shot setting also performed well.
Team OpenFact [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] finetuned DeBERTa and mDeBERTa on multiple versions of the task dataset. This
included training one model per language using the corresponding language train subset. The team
also experimented with multilingual models by training over concatenated train subsets of all (or part)
of the task four languages.
      </p>
      <p>
        Team HYBRINFOX [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] developed a classification pipeline, consisting of three parts: a standard
language model (RoBERTa for English and multilingual BERT for other languages), a component for
extracting and encoding triples using OpenIE6 and Multi2OIE, and a merging neural network with a
softmax layer for output. Early results indicated that including the triple encoding component improved
performance over using the language model alone, especially for English. Challenges were noted in
evaluating the approach for Dutch and Arabic due to limited proficiency in these languages.
Team DSHacker [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] conducted experiments with both monolingual and multilingual approaches. For
the monolingual approach, BERT models were fine-tuned for specific languages. For the multilingual
approach, XLM-RoBERTa-large was used, initially optimized and fine-tuned on the entire dataset.
In a subsequent experiment, Spanish was excluded from the training data. Additionally, two LLMs,
GPT-3.5-turbo and the recently released GPT-4o, were employed for each language using few-shot
prompting to classify texts. A model was also fine-tuned on the DIPROMATS 2024 Task 1 dataset to
predict whether the data from CheckThat! Lab 2024 Task 1 contained propaganda. This analysis aimed
to indirectly determine whether check-worthy data also included propaganda. The XLM-RoBERTa-large
model, fine-tuned for binary propaganda classification, was further fine-tuned for check-worthiness
classification.
      </p>
      <p>
        Team FC_RUG [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] tested GEITje, an LLM for Dutch based on Mistral-7B. They experimented with
diferent prompts varying the learning settings (zero-shot vs few-shot) and the personas (helpful
assistant vs fact-checker). The best model with few-shot in-context learning was selected based on the
development data from the companion task of the CheckThat! 2022 Lab edition.
      </p>
      <p>
        Team CLaC [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] approached the task as a binary classification task, leveraging a LLM (Google’s Gemini 4)
to classify whether a sentence is True or False, without specifying the task to classify for. The task
was modeled as a multi-annotator scenario where Gemini was used to create two semantically-similar
sentences to each test sentence. Then, Gemini was prompted to predict one of these labels: True,
or False, for each sentence, using a single prompt. Finally, majority vote over the three annotations
was used as the final label. Additionally, to improve performance, the prompt was contextualized by
providing 600 randomly selected samples from the training subset.
      </p>
      <p>
        Team SINAI [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] attempted two diferent approaches were attempted: (i) RoBERTa-base was fine-tuned
using the original English data, and data augmentation was tried with Spanish transcription-sourced
texts; (ii) A prompting approach with GPT-3.5-turbo was conducted, involving two experiments: one
concatenating previous consecutive examples from the data (using the sentence_id) and the other
using only the original text. Finally, after analyzing the results obtained from both approaches, the
RoBERTa-base fine-tuning approach with the original English data was elected.
      </p>
      <p>
        Team Trio Titans [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] fine-tuned diferent transformer models including DistilBERT, ALBERT, and
RoBERTa, with the latter performing the best.
      </p>
      <p>
        Team DataBees [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] fine-tuned various pre-trained models such as BERT, RoBERTa, and
languagespecific models like AraBERT for Arabic, along with traditional classifiers like MultinomialNB and
Logistic Regression. The system was designed to work across the three languages. Their best F1 scores
were achieved with DistilBERT for English, AraBERT for Arabic, and MultinomialNB for Dutch.
Team TurQUaz [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] developed difernet models for each language. For Arabic and English, a two-stage
approach was proposed to determine check-worthy statements. This method combined a fine-tuned
RoBERTa classifier with in-context learning (ICL) using multiple diferent instruct-tuned models. The
aggregation method varied between the Arabic and English datasets. For the Dutch dataset, the
ifne-tuned classifier was excluded, and reliance was placed solely on in-context learning due to time
constraints.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Related Work</title>
      <sec id="sec-5-1">
        <title>5.1. Checkworiness in Fact-checking</title>
        <p>
          Due to the significant surge of disinformative content online the importance of improving the capabilities
of fact-checking pipeline is paramount. As depicted in Figure 1, the first part of the pipeline is finding
claims that important to fact check [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. The overall idea is to facilitate human fact-checkers to
seamlessly streamline their daily fact-checking activities. To address and improve the capabilities
of diferent components of fact-checking pipeline, there has been a considerable surge in research
consisting of exploring fact-checking perspectives on fake news and associated issues [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], examining
attitudes towards the detection of misinformation and disinformation [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ], automating fact-checking to
support human fact-checkers [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], predicting the factuality and the bias of entire news outlets [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ],
detecting disinformation across multiple modalities [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ], and focusing on the use of abusive language
on social media [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ].
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. LLMs for Checkworthiness Task</title>
        <p>
          Given that large language models (LLMs) have been demonstrating significant capabilities across
various disciplines and many downstream NLP tasks, eforts have been made to utilize such models for
detecting claims and their worthiness. Majer and Šnajder [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] evaluated gpt-4-turbo and demonstrated
its potential for claim check-worthiness detection with minimal prompt engineering. Sawiński et al.
[
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] used GPT-3.5 and GPT-4 models in zero-shot and few-shot learning setups, comparing them with
GPT-3, BERT, and RoBERTa-based fine-tuned models. Their findings demonstrate that the fine-tuned
GPT-3 model performed the best across diferent models. Abdelali et al. [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ] benchmarked various
open and closed models for the Arabic checkworthiness task using the CT–CWT–22 dataset [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ] and
demonstrated that the performance of few-shot learning using GPT-4 is relatively higher; however, it is
still far from state-of-the-art performance.
        </p>
        <p>
          CT! Lab
CT-2018 [40] Debate Text Ar, En
CT-2019 [41] Debate, Web pages Text Ar, En
CT-2020 [42] Tweet Text Ar, En
CT-2021 [43, 44] Tweet, debate Text Ar, Bg, En, Es, Tr
CT-2022 [
          <xref ref-type="bibr" rid="ref35">35, 45</xref>
          ] Tweet Text Ar, Bg, En, Nl, Es, Tr
CT-2023 [46, 47] Tweet Text, Image Ar, En
CT-2024 Tweet, debate Text Ar, En, Nl
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Previous Editions of Checkworthiness Shared Tasks</title>
        <p>Since the seminal work by Hassan et al. [36], the task of check-worthiness estimation has gained broader
interest. This task, proposed by Hassan et al. [36], involves assessing whether a sentence from a political
debate is non-factual, trivially factual, or significantly factual enough to warrant verification. Since then,
several notable studies have focused on political debates [37], tweets, and transcripts from political
debates [38], as well as cross-lingual studies over tweets [39].</p>
        <p>A major research interest has been sparked since the inception of the CLEF CheckThat!lab initiatives.
The initial focus was primarily on political debates and speeches. This focus has since expanded to
include social media, transcriptions, and various languages and modalities.</p>
        <p>Significant research interest has been sparked since the inception of the CLEF CheckThat!lab
initiatives. The initial focus was primarily on political debates and speeches. This focus has since
expanded to include social media, transcriptions, and various languages and modalities. In Table 4, we
report a summary of check-worthiness tasks over the years from 2018 to 2024. The focus has mainly
been on debates and tweets, mostly in the text modality. As for languages, Arabic and English have
been ofered in all editions. The number of participants and system description paper submissions has
increased over the years.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and Future Work</title>
      <p>We presented an overview of Task 1 of the CLEF-2024 CheckThat! lab, which focused on
checkworthiness estimation of multigenre content and covering three languages: Arabic, Dutch, and English.
The task attracted significant participation, with 75 registered teams and 28 teams submitting system
description papers. The majority of the participating systems leveraged transformer-based models,
showcasing their efectiveness in this domain. Notable approaches included the fine-tuning of
languagespecific models such as AraBERT for Arabic and RobBERT for Dutch, as well as the use of multilingual
models like XLM-RoBERTa. Several teams experimented with large language models including
GPT3.5 and Llama2, while others implemented ensemble approaches combining multiple models. Data
augmentation and preprocessing techniques were widely employed to enhance performance, and some
teams incorporated named entity recognition and other linguistic features into their systems. The
results show significant improvements over the baselines across all languages, highlighting the progress
made in check-worthiness estimation. Future work may include covering other modalities and domains.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>The work of F. Alam, M. Hasanain, R. Suwaileh and W. Zaghouani is partially supported by NPRP
14C0916-210015 from the Qatar National Research Fund, which is a part of Qatar Research Development and
Innovation Council (QRDI). The findings achieved herein are solely the responsibility of the authors.
the Evaluation Forum, CLEF ’2022, Bologna, Italy, 2022.
[36] N. Hassan, C. Li, M. Tremayne, Detecting check-worthy factual claims in presidential debates,
in: Proceedings of the 24th ACM International on Conference on Information and Knowledge
Management, CIKM ’15, 2015, pp. 1835–1838.
[37] S. Vasileva, P. Atanasova, L. Màrquez, A. Barrón-Cedeño, P. Nakov, It takes nine to smell a rat:
Neural multi-task learning for check-worthiness prediction, in: Proceedings of the International
Conference on Recent Advances in Natural Language Processing, RANLP ’19, 2019, pp. 1229–1239.
[38] Y. S. Kartal, M. Kutlu, Re-think before you share: A comprehensive study on prioritizing
checkworthy claims, IEEE Transactions on Computational Social Systems (2022).
[39] M. Hasanain, T. Elsayed, Cross-lingual transfer learning for check-worthy claim identification
over twitter, arXiv: 2211.05087 (2022).
[40] P. Atanasova, L. Marquez, A. Barrón-Cedeño, T. Elsayed, R. Suwaileh, W. Zaghouani, S. Kyuchukov,
G. Da San Martino, P. Nakov, Overview of the CLEF-2018 CheckThat! lab on automatic
identification and verification of political claims. Task 1: Check-worthiness, CEUR Workshop Proceedings,
2018.
[41] P. Atanasova, P. Nakov, G. Karadzhov, M. Mohtarami, G. Da San Martino, Overview of the
CLEF-2019 CheckThat! lab on automatic identification and verification of claims. Task 1:
Checkworthiness, CEUR Workshop Proceedings, 2019.
[42] A. Barrón-Cedeño, T. Elsayed, P. Nakov, G. Da San Martino, M. Hasanain, R. Suwaileh, F. Haouari,
N. Babulkov, B. Hamdan, A. Nikolov, S. Shaar, Z. Sheikh Ali, Overview of CheckThat! 2020:
Automatic identification and verification of claims in social media, in: A. Arampatzis, E. Kanoulas,
T. Tsikrika, S. Vrochidis, H. Joho, C. Lioma, C. Eickhof, A. Névéol, L. Cappellato, N. Ferro (Eds.),
Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Eleventh
International Conference of the CLEF Association (CLEF 2020), LNCS (12260), Springer, 2020, pp.
215–236.
[43] S. Shaar, M. Hasanain, B. Hamdan, Z. S. Ali, F. Haouari, A. Nikolov, M. Kutlu, Y. S. Kartal, F. Alam,
G. Da San Martino, A. Barrón-Cedeño, R. Míguez, J. Beltrán, T. Elsayed, P. Nakov, Overview of the
CLEF-2021 CheckThat! lab task 1 on check-worthiness estimation in tweets and political debates,
2021.
[44] P. Nakov, G. Da San Martino, T. Elsayed, A. Barrón-Cedeño, R. Míguez, S. Shaar, F. Alam, F. Haouari,
M. Hasanain, W. Mansour, B. Hamdan, Z. S. Ali, N. Babulkov, A. Nikolov, G. K. Shahi, J. M. Struß,
T. Mandl, M. Kutlu, Y. S. Kartal, Overview of the CLEF-2021 CheckThat! lab on detecting
check-worthy claims, previously fact-checked claims, and fake news, in: K. Candan, B. Ionescu,
L. Goeuriot, B. Larsen, H. Müller, A. Joly, M. Maistro, F. Piroi, G. Faggioli, N. Ferro (Eds.),
Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Twelfth
International Conference of the CLEF Association, LNCS (12880), 2021.
[45] P. Nakov, A. Barrón-Cedeño, G. Da San Martino, F. Alam, J. M. Struß, T. Mandl, R. Míguez, T. Caselli,
M. Kutlu, W. Zaghouani, C. Li, S. Shaar, G. K. Shahi, H. Mubarak, A. Nikolov, N. Babulkov, Y. S.
Kartal, J. Beltrán, M. Wiegand, M. Siegel, J. Köhler, Overview of the CLEF-2022 CheckThat! lab on
ifghting the COVID-19 infodemic and fake news detection, in: A. Barrón-Cedeño, G. Da San
Martino, M. Degli Esposti, F. Sebastiani, C. Macdonald, G. Pasi, A. Hanbury, M. Potthast, G. Faggioli,
F. Nicola (Eds.), Proceedings of the 13th International Conference of the CLEF Association:
Information Access Evaluation meets Multilinguality, Multimodality, and Visualization, CLEF ’2022,
Bologna, Italy, 2022.
[46] F. Alam, A. Barrón-Cedeño, G. S. Cheema, S. Hakimov, M. Hasanain, C. Li, R. Míguez, H. Mubarak,
G. K. Shahi, W. Zaghouani, P. Nakov, Overview of the CLEF-2023 CheckThat! lab task 1 on
check-worthiness in multimodal and multigenre content, in: M. Aliannejadi, G. Faggioli, N. Ferro,
Vlachos, Michalis (Eds.), Working Notes of CLEF 2023–Conference and Labs of the Evaluation
Forum, CLEF 2023, Thessaloniki, Greece, 2023.
[47] A. Barrón-Cedeño, F. Alam, T. Caselli, G. Da San Martino, T. Elsayed, A. Galassi, F. Haouari,
F. Ruggeri, J. M. Struß, R. N. Nandi, G. S. Cheema, D. Azizov, P. Nakov, The CLEF-2023 CheckThat!
Lab: Checkworthiness, subjectivity, political bias, factuality, and authority, in: J. Kamps, L. Goeuriot,
F. Crestani, M. Maistro, H. Joho, B. Davis, C. Gurrin, U. Kruschwitz, A. Caputo (Eds.), Advances in
Information Retrieval, Springer Nature Switzerland, Cham, 2023, pp. 506–517.
[48] G. Faggioli, N. Ferro, P. Galuščáková, A. García Seco de Herrera (Eds.), Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, CLEF 2024, 2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <surname>The</surname>
            <given-names>CLEF</given-names>
          </string-name>
          -2024 CheckThat! Lab:
          <article-title>Check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Advances in Information Retrieval</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>449</fpage>
          -
          <lpage>458</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Piskorski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2024 CheckThat! Lab: Check-worthiness, subjectivity, persuasion, roles, authorities and adversarial robustness</article-title>
          , in: L.
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mulhem</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Quénot</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Soulier</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Di Nunzio</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Galuščáková</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ),
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dalvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sajjad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D. S.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abdelali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Durrani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Darwish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Homaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Danoe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Stolk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bruntink</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>Fighting the COVID-19 infodemic: Modeling the perspective of journalists, fact-checkers, social media platforms, policy makers, and the society</article-title>
          ,
          <source>in: Findings of EMNLP</source>
          <year>2021</year>
          ,
          <year>2021</year>
          , pp.
          <fpage>611</fpage>
          -
          <lpage>649</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Arslan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tremayne</surname>
          </string-name>
          ,
          <article-title>A benchmark dataset of check-worthy factual claims</article-title>
          ,
          <source>in: Proceedings of the International AAAI Conference on Web and Social Media</source>
          , volume
          <volume>14</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>821</fpage>
          -
          <lpage>829</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Chandani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. E. Z.</given-names>
            <surname>Syeda</surname>
          </string-name>
          , Checker Hacker at CheckThat! 2024:
          <article-title>Ensemble models for checkworthy tweet identification</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gruman</surname>
          </string-name>
          , L. Kosseim, CLaC at CheckThat!
          <year>2024</year>
          :
          <article-title>A zero-shot model for check-worthiness and subjectivity classification</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Sriram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Venkatesh</surname>
          </string-name>
          , Databees at CheckThat! 2024:
          <article-title>Check worthiness estimation</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Golik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Modzelewski</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Jochym, DSHacker at CheckThat! 2024:
          <article-title>LLMs and BERT for check-worthy claims detection with propaganda co-occurrence analysis</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Panchendrarajan</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Zubiaga, FactFinders at CheckThat! 2024:
          <article-title>Refining check-worthy statement detection with LLMs through data pruning</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Weering</surname>
          </string-name>
          , T. Caselli, FC_RUG at CheckThat! 2024:
          <article-title>Few-shot learning using GEITje for checkworthiness detection in Dutch</article-title>
          , in: [48],
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M. S. A.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Shanto</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Chowdhury</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Murad</surname>
          </string-name>
          , U. Das,
          <article-title>Fired_from_NLP at CheckThat! 2024: Estimating the check-worthiness of tweets using a fine-tuned transformer-based approach</article-title>
          , in: [48],
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>I.</given-names>
            <surname>Vogel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Möhle</surname>
          </string-name>
          , Fraunhofer SIT at CheckThat! 2024:
          <article-title>Adapter fusion for check-worthiness detection</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Faye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Casanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chanson</surname>
          </string-name>
          , G. Gadek, G. Gravier, P. Égré, HYBRINFOX at CheckThat! 2024:
          <article-title>Enhancing language models with structured information for checkworthiness estimation</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P. R.</given-names>
            <surname>Aarnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          , P. Galuščáková, IAI group at CheckThat! 2024:
          <article-title>Transformer models and data augmentation for checkworthy claim detection</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>M. Dryankova1</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Dimitrov</surname>
            ,
            <given-names>I. Koychev</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          , Mirela at CheckThat! 2024:
          <article-title>Check-worthiness of tweets with multilingual embeddings and adversarial training</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sawinski</surname>
          </string-name>
          , OpenFact at CheckThat! 2024:
          <article-title>Optimizing training data selection through undersampling techniques</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A. I.</given-names>
            <surname>Paran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Shohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ahsan</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hoque</surname>
          </string-name>
          , SemanticCuetSync at CheckThat! 2024:
          <article-title>Finding subjectivity in news article using Llama</article-title>
          , in: [48],
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J. Valle</given-names>
            <surname>Aguilera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J. Gutiérrez</given-names>
            <surname>Megías</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Jiménez Zafra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Ureña López</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Martínez Cámara</surname>
          </string-name>
          , SINAI at CheckThat! 2024:
          <article-title>Stealthy character-level adversarial attacks using homoglyphs and search, iterative</article-title>
          , in: [48],
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S. B. K.</given-names>
            <surname>Giridharan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sounderrajan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. R.</given-names>
            <surname>Salim</surname>
          </string-name>
          , SSN-NLP at CheckThat! 2024:
          <article-title>Assessing the check-worthiness of tweets and debate excerpts using traditional machine learning and transformer models</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Prarthna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. V.</given-names>
            <surname>Chiranjeev Prasannaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Sai</given-names>
            <surname>Geetha</surname>
          </string-name>
          , Trio Titans at CheckThat! 2024:
          <article-title>Check worthiness estimation</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Bulut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. E.</given-names>
            <surname>Keleş</surname>
          </string-name>
          , M. Kutlu, TurQUaz at CheckThat!
          <year>2024</year>
          :
          <article-title>A hybrid approach of fine-tuning and in-context learning for check-worthiness estimation</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Shohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. I.</given-names>
            <surname>Paran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hoque</surname>
          </string-name>
          , SemanticCuetSync at CheckThat! 2024:
          <article-title>Finetuning transformer models for checkworthy tweet identification</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Weering</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2024 CheckThat! lab task 1 on check-worthiness estimation of multigenre content</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>S.</given-names>
            <surname>Stoia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Montañez-Collado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ibáñez-Bautista</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montejo-Ráez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Martín-Valdivia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Díaz-Galiano</surname>
          </string-name>
          , SINAI at CheckThat! 2024:
          <article-title>Transformer-based approaches for check-worthiness classification</article-title>
          ,
          <source>in: [48]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Corney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          , G. Da San Martino,
          <article-title>Automated fact-checking for assisting human fact-checkers</article-title>
          ,
          <source>in: Proceedings of the 30th International Joint Conference on Artificial Intelligence, IJCAI '21</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>4551</fpage>
          -
          <lpage>4558</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          , Automated fact checking:
          <article-title>Task formulations, methods and future directions</article-title>
          ,
          <source>in: Proceedings of the 27th International Conference on Computational Linguistics</source>
          , COLING '18,
          <string-name>
            <surname>Association</surname>
            for Computational Linguistics, Santa Fe,
            <given-names>NM</given-names>
          </string-name>
          , USA,
          <year>2018</year>
          , pp.
          <fpage>3346</fpage>
          -
          <lpage>3359</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Arora</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Augenstein,</surname>
          </string-name>
          <article-title>A survey on stance detection for mis-and disinformation identification</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: NAACL</source>
          <year>2022</year>
          ,
          <year>2022</year>
          , pp.
          <fpage>1259</fpage>
          -
          <lpage>1277</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <article-title>Fakekg: A knowledge graph of fake claims for improving automated fact-checking (student abstract)</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>37</volume>
          (
          <year>2023</year>
          )
          <fpage>16320</fpage>
          -
          <lpage>16321</lpage>
          . doi:
          <volume>10</volume>
          .1609/aaai.v37i13.
          <fpage>27020</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. T.</given-names>
            <surname>Sencar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kwak</surname>
          </string-name>
          ,
          <article-title>A survey on predicting the factuality and the bias of news media</article-title>
          ,
          <source>arXiv/2103</source>
          .12506 (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cresci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          , G. Da San Martino, S. Shaar,
          <string-name>
            <given-names>H.</given-names>
            <surname>Firooz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>A survey on multimodal disinformation detection</article-title>
          ,
          <source>in: Proceedings of the 29th International Conference on Computational Linguistics</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>6625</fpage>
          -
          <lpage>6643</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>A.</given-names>
            <surname>Arora</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Sarwar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Nayak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dinkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zlatkova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhatawdekar</surname>
          </string-name>
          , G. Bouchard,
          <string-name>
            <surname>I. Augenstein</surname>
          </string-name>
          ,
          <article-title>Detecting harmful content on online platforms: What platforms need vs. where research eforts go</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>5</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>L.</given-names>
            <surname>Majer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Šnajder</surname>
          </string-name>
          ,
          <article-title>Claim check-worthiness detection: How well do llms grasp annotation guidelines?</article-title>
          ,
          <source>arXiv:2404.12174</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sawiński</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Węcel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. P.</given-names>
            <surname>Księżniak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Stróżyna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lewoniewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Stolarski</surname>
          </string-name>
          , W. Abramowicz, Openfact at checkthat!
          <year>2023</year>
          <article-title>: head-to-head gpt vs. bert-a comparative study of transformers language models for the detection of check-worthy claims</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3497</volume>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>A.</given-names>
            <surname>Abdelali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mousi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Boughorbel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abdaljalil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>El Kheir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Izham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dalvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hawasly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nazar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Elshahawy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Durrani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>MilicFrayling</surname>
          </string-name>
          , F. Alam,
          <article-title>LAraBench: Benchmarking Arabic AI with large language models</article-title>
          , in: Y. Graham, M. Purver (Eds.),
          <source>Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics, St. Julian's, Malta</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>487</fpage>
          -
          <lpage>520</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beltrán</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2022 CheckThat! lab task 1 on identifying relevant claims in tweets</article-title>
          , in: N. Faggioli, Guglielmo andd Ferro,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          , M. Potthast (Eds.),
          <source>Working Notes of CLEF 2022-Conference and Labs of</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>