<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Y. Liu, J. Gu, N. Goyal, X. Li, S. Edunov, M. Ghazvininejad, M. Lewis, L. Zettlemoyer, Multi-
lingual denoising pre-training for neural machine translation, Transactions of the Association
for Computational Linguistics</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1162/tacl_a_00343</article-id>
      <title-group>
        <article-title>at CheckThat! 2025: Switching Fine-Tuned SLMs and LLM Prompting for Multilingual Claim Normalization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fabrycio Leite Nakano Almada</string-name>
          <email>fabrycio@egresso.ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kauan Divino Pouso Mariano</string-name>
          <email>kauan@discente.ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maykon Adriell Dutra</string-name>
          <email>maykonadriell@discente.ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Emanuel da Silva Monteiro</string-name>
          <email>victor_emanuel@discente.ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juliana Resplande Sant' Anna Gomes</string-name>
          <email>juliana.resplande@discente.ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arlindo Rodrigues Galvão Filho</string-name>
          <email>arlindogalvao@ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anderson da Silva Soares</string-name>
          <email>andersonsoares@ufg.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Claim Normalization, Disinformation, Multilingual NLP, Fact-Checking, Transformer Models, Zero-Shot Learning</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Advanced Knowledge Center in Immersive Technology (AKCIT), Federal University of Goiás</institution>
          ,
          <country country="BR">Brazil</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Informatics, Federal University of Goiás</institution>
          ,
          <country country="BR">Brazil</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>8</volume>
      <issue>2020</issue>
      <fpage>0000</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>Claim normalization, the transformation of informal social media posts into concise, self-contained statements, is a crucial step in automated fact-checking pipelines. This paper details our submission to the CLEF-2025 CheckThat! Task 2, which challenges systems to perform claim normalization across twenty languages, divided into thirteen supervised (high-resource) and seven zero-shot (no training data) tracks. Our approach, leveraging fine-tuned Small Language Models (SLMs) for supervised languages and Large Language Model (LLM) prompting for zero-shot scenarios, achieved podium positions (top three) in fiteen of the twenty languages. Notably, this included second-place rankings in eight languages, five of which were among the seven designated zero-shot languages, underscoring the efectiveness of our LLM-based zero-shot strategy . For Portuguese, our initial development language, our system achieved an average METEOR score of 0.5290, ranking third. All implementation artifacts, including inference, training, evaluation scripts, and prompt configurations , are publicly available at https://github.com/ju-resplande/checkthat2025_normalization.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The proliferation of misinformation within social media ecosystems has intensified the demand for
automated fact-checking pipelines that operate efectively across diverse languages , genres, and text
characterized by significant noise . A crucial stage in such pipelines is claim normalization: the
process of converting an informal, often multi-sentence social media post into a concise , self-contained
statement suitable for subsequent evidence retrieval and veracity assessment. Without normalization,
downstream modules must contend with extraneous elements such as redundancy, hashtags, emojis,
and idiosyncratic phrasing, which collectively diminish both retrieval recall and factual accuracy [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>The CLEF-2025 CheckThat! Lab confronts this bottleneck through Task 2 – Claim Normalization.
Systems are required to generate normalized claims for twenty languages under two experimental
conditions: (i) a monolingual setting, providing training and development splits (annotated examples)</p>
      <p>CEUR</p>
      <p>
        ceur-ws.org
for thirteen higher-resource languages, and (ii) a zero-shot setting, releasing only test data for seven
lower-resource languages, for which no specific training data is provided [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ].
      </p>
      <p>Our methodological development initially focused on Portuguese—the native language of our
development team, facilitating a more nuanced understanding and iterative refinement —before extending
validated strategies to all target languages. For languages in category (i), our experimentation involved
ifne-tuning open-source Encoder-Decoder Small Language Models (SLMs) and, as a comparative
approach, inference via prompting with Large Language Models (LLMs). For languages in category (ii),
we exclusively employed a zero-shot prompting strategy with LLMs.</p>
      <p>Our eforts culminated in strong results , securing third place in the Portuguese subset with an average
METEOR score of 0.5290. Across all languages, we achieved top-three placements in fiteen of the
twenty languages. This included second-place finishes in eight languages overall (with five of these
being zero-shot languages) and third-place finishes in seven languages , highlighting the robustness of
our approach in both data-rich and data-scarce scenarios.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Over the past few years, claim normalization has gained prominence as a crucial preprocessing step
in automated fact-checking, moving beyond mere claim extraction. Konstantinovskiy et al. proposed
the first annotation schema for claim detection informed by experts and a benchmark for automated
claim detection that is more consistent across time, topics, and annotators than previous approaches.
Sundriyal et al. formalized claim normalization by converting informal social-media texts into
selfcontained claims, demonstrating significant improvements in downstream evidence retrieval and
veracity classification tasks .</p>
      <p>
        Previous editions of the CLEF CheckThat! Lab (2018–2024) have included tasks such as
checkworthiness detection from political debates or speeches [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 7, 8, 9</xref>
        ], and from political or
COVID-19related tweets [
        <xref ref-type="bibr" rid="ref8">8, 9</xref>
        ].
      </p>
      <p>The construction of multilingual fact-checking corpora represents another active research direction.
For instance, MuMiN automatically links 21 million tweets to 13 thousand fact-checked claims across
41 languages using LaBSE embeddings [10]. Similarly, Singh et al. created MMTweets by scraping
”debunked narratives,” retrieving tweets via multilingual keyword queries, and applying detailed human
annotation, resulting in a multimodal dataset for cross-lingual retrieval.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Task Definition</title>
      <p>
        Task 2 of the CLEF-2025 CheckThat! Lab presents a comprehensive multilingual claim normalization
challenge designed to evaluate system performance across diverse linguistic contexts and resource
availability scenarios. The core objective involves transforming informal social media posts into
concise, self-contained, and verifiable statements while preserving all factual content and systematically
removing subjective opinions, redundant expressions, and extraneous material [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ].
      </p>
      <p>The task employs two experimental paradigms designed to assess system robustness under varying
data availability conditions:
a. Monolingual setting: Complete training, development, and test datasets are provided for
thirteen languages: Arabic (AR), English (EN), French (FR), German (DE), Hindi (HI), Indonesian
(ID), Marathi (MR), Polish (PL), Portuguese (PT), Punjabi (PA), Spanish (ES), Tamil (TA), and
Thai (TH).
b. Zero-shot setting: For seven languages, Only test splits are released for seven languages,
requiring systems to generalize from cross-lingual knowledge or leverage multilingual
pretraining: Bengali (BN), Czech (CS), Dutch (NL), Greek (EL), Korean (KO), Romanian (RO), and
Telugu (TE).</p>
      <sec id="sec-3-1">
        <title>I guess the left is okay with this I guess the left is okay</title>
        <p>with this I guess the left is okay with this Dr . Rachel</p>
      </sec>
      <sec id="sec-3-2">
        <title>Levine @DrRachel Levine Thank you Vanity Fair for hon</title>
        <p>oring me on the cover of your magazine this March. My
dream of becoming @POTUS one day just took a step
forward. THE SKY THE LIMIT. ”Madam President Levine</p>
      </sec>
      <sec id="sec-3-3">
        <title>A LEADER IN THE MAKING 8:12 AM Feb 1, 2021 Twitter Web App ... .</title>
        <sec id="sec-3-3-1">
          <title>Normalized Claim</title>
        </sec>
      </sec>
      <sec id="sec-3-4">
        <title>US assistant health secretary Rachel</title>
      </sec>
      <sec id="sec-3-5">
        <title>Levine appears on the cover of Vanity Fair’s</title>
      </sec>
      <sec id="sec-3-6">
        <title>March 2021 issue</title>
        <p>Participant submissions were evaluated using the METEOR score [12], a metric commonly employed
for machine translation evaluation that assesses translation quality by aligning system output with
reference texts based on exact word matches, stemming, and synonymy. For this task, oficial scores
were calculated by averaging METEOR results across all test examples for a given language/setting.
Punctuation was removed during pre-processing for the evaluation script to standardize inputs and
mitigate its impact on scores.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>To address the claim normalization task, we employed a dual-strategy approach tailored to the diferent
experimental settings. For the monolingual setting, where training, development, and test splits were
available, we investigated both: (i) fine-tuning of open-source Encoder-Decoder Small Language Models
(SLMs), and (ii) inference using Large Language Models (LLMs) with few-shot prompting. For the
zero-shot setting, characterized by the absence of training data for specific languages , our eforts
exclusively focused on zero-shot inference with LLMs (using prompts without in-context examples
specific to the task for those languages ).</p>
      <p>Initial experimentation, including preliminary model selection, hyperparameter tuning, and
qualitative analysis, was conducted on the Portuguese dataset. This choice was motivated by the team’s
native proficiency in the language , facilitating a more nuanced understanding of model behavior and
output quality before extending the approach to other languages.</p>
      <p>The subsequent subsections detail our data cleaning pipeline, exploratory data analysis insights, the
specifics of our modeling techniques , and the evaluation framework.
4.1. Data Cleaning
A recurrent issue observed across multiple languages was the presence of triplicated sentences within
original posts, often appended with a None placeholder. This pattern, likely an artifact of automated
data collection or formatting, is exemplified in Table 3 for a Portuguese instance.</p>
      <sec id="sec-4-1">
        <title>Post Original: Na Holanda, a ministra da Saúde trabalha</title>
        <p>duas (2) horas diariamente como agente de limpeza antes de
ir ao seu escritório. Gostei muito. Na Holanda, a ministra
da Saúde trabalha duas (2) horas diariamente como agente
de limpeza antes de ir ao seu escritório. Gostei muito. Na</p>
      </sec>
      <sec id="sec-4-2">
        <title>Holanda, a ministra da Saúde trabalha duas (2) horas diari</title>
        <p>amente como agente de limpeza antes de ir ao seu escritório.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Gostei muito. None</title>
      </sec>
      <sec id="sec-4-4">
        <title>Saída Normalizada: Na Holanda, a ministra da Saúde</title>
        <p>trabalha duas horas diariamente como agente de limpeza
antes de ir ao seu escritório.</p>
        <p>English Translation</p>
      </sec>
      <sec id="sec-4-5">
        <title>Original Post: In the Netherlands, the Minister of Health</title>
        <p>works two (2) hours daily as a cleaning worker before
going to her office. I really liked that. In the Netherlands, the</p>
      </sec>
      <sec id="sec-4-6">
        <title>Minister of Health works two (2) hours daily as a cleaning</title>
        <p>worker before going to her ofice . I really liked that. In the</p>
      </sec>
      <sec id="sec-4-7">
        <title>Netherlands, the Minister of Health works two (2) hours daily as a cleaning worker before going to her ofice . I really liked that. None</title>
      </sec>
      <sec id="sec-4-8">
        <title>Normalized Claim: In the Netherlands, the Minister of</title>
        <p>Health works two hours daily as a cleaning worker before
going to her ofice .</p>
        <p>To rectify this, we implemented a preprocessing routine (Algorithm 1) designed to first remove any
trailing None tokens. Subsequently, it identifies and condenses repeated textual sequences by searching
for the smallest repeating pattern that constitutes the entire post. If such a pattern is found, only a
single instance of it is retained.</p>
        <p>Algorithm 1 Preprocessing for Repetitive Content and Placeholder Removal
Require: Raw post 
Ensure: Cleaned post  
 clean ← 
if  ends with “None” then</p>
        <p>clean ←  clean[∶ −4].strip()
end if
 words ← tokenize( clean)
for  = 1 to ⌊| words|/2⌋ do
 ←  words[0 ∶ ]
num_repeats ← ⌊| words|/⌋
repeated_sequence ←  repeated num_repeats times
if  words[0 ∶  ⋅ num_repeats] = repeated_sequence then
 clean ← join( , spaces)
return  clean
end if
end for
return  clean
▷ Remove trailing placeholder</p>
        <p>▷ Tokenize into words
▷ Check for repetitive patterns
▷ Extract candidate pattern
▷ Return single instance
▷ No repetition found</p>
        <p>Additionally, we implemented cross-split deduplication to handle identical posts appearing in
multiple dataset partitions. To preserve evaluation integrity, duplicates were systematically removed by
prioritizing retention in test sets, then development sets, and finally training sets .
4.2. Exploratory Data Analysis (EDA)
We analyzed word count distributions for original posts and their corresponding normalized claims
within the Portuguese subset following preprocessing (Algorithm 1). As depicted for the training and
development sets in Figures 1a and 1b, original posts (mean ≈ 75 words, STD ≈ 107 words; highly
skewed) are substantially longer and more variable than normalized claims (mean ≈ 15 words, STD ≈
6.8 words). This demonstrates that normalization efectively reduces verbosity and structural noise ,
leading to more compact and verifiable statements .</p>
        <p>Train
Dev
Median = 41.0
Train
Dev
Median = 14.0
1.60%
1.40%
1.20%
e1.00%
g
a
t
en0.80%
c
r
e
P0.60%
0.40%
0.20%
0.00%
0
500 1000</p>
        <p>Word count
1500
0
20</p>
        <p>40
Word count
60
80
(a) Original posts word count (Portuguese).</p>
        <p>(b) Normalized claims word count (Portuguese).</p>
        <p>The dataset’s heterogeneity across languages, as indicated in Table 2, presents further challenges.
While some languages ofer substantial training samples , others, particularly those in the zero-shot
group, provide only minimal test data. This disparity necessitates careful consideration of model
generalization and poses a risk of overfitting in resource-rich scenarios and underperformance in
low-resource ones.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments</title>
      <p>Our experimental design directly follows the dual-strategy approach detailed in Section 4, addressing
both monolingual (with training and development data) and zero-shot (test data only for specific
languages) settings. All input posts were preprocessed according to Algorithm 1.
5.1. Fine-tuning of Encoder-Decoder SLMs
This approach involved adapting existing pre-trained encoder-decoder Transformer models for the
specific task of claim normalization , utilizing the available training data for the thirteen supervised
languages. We primarily sourced these models from the Hugging Face Hub1. Our strategy prioritized
monolingual models, selecting those pre-trained specifically for each target language . The selected
monolingual models included:
• Portuguese: PTT5 (small, base, large) [13], PTT5-mMARCO (base) [14], PTT5-v2 (small, base,
large) [15], Mono-PTT5 (small, base, large) [15], Portuguese Bart (base) [16].
• Arabic: AraT5 (base) [17].
• French: T5 French (base) [18].
• German: T5 German (small) [19].
• Indic languages - Hindi, Marathi, Punjabi, Tamil: Varta T5 (base) [20].
• Indonesian: Indonesian T5 Summarization Base [21].
• Polish: PLT5 (base) [22].
• Spanish: T5S (base) [23].</p>
      <p>• Thai: ThaiT5 Instruct (base) [24].</p>
      <p>In addition to these language-specific models , we also experimented with fine-tuning the following
multilingual encoder-decoder architectures on the combined training data of all supervised languages:
Flan-T5 (small, base, large) [14], mBART (large) [25], and UMT5 (base) [26].</p>
      <p>Fine-tuning was conducted on three distinct hardware platforms: Kaggle kernels equipped with
two NVIDIA T4 GPUs, Google Colab Pro sessions with a single NVIDIA T4 GPU, and an on-premise
server hosting a single NVIDIA A100 GPU. The A100 GPU, with its substantial memory capacity, was
crucial for fine-tuning larger models like PTT5 (Large), which would exceed the memory limits of the
T4 GPUs.</p>
      <p>We used hyperparameter search space, summarized in Table 4, in which a gradient accumulation was
employed to achieve an efective batch size of 32 , even on GPUs with 16 GB of VRAM. The maximum
generation length for each batch was dynamically set to the length of the longest target sequence in
that batch plus two tokens, minimizing unnecessary padding.
5.2. Inference with LLMs
For the monolingual setting, as an alternative to fine-tuning SLMs , we investigated the capabilities of
LLMs using few-shot in-context learning in two distinct scenarios: (1) few-shot in-context learning as
an alternative to fine-tuning smaller language models (SLMs) in monolingual settings, and (2) zero-shot
inference for cross-lingual transfer to languages without available training data. The following LLMs
were experimented via their respective APIs:
• Google Gemini [27]: Gemini 2.0 Flash Lite, Gemini 2.0 Flash Thinking.
• OpenAI GPT [28]: GPT-4o, GPT-4o mini, GPT-4.1 mini.
• OpenAI Reasoning [29, 30]: o1, o3 mini.
• Mistral Pixtral2: Pixtral Large (124B).</p>
      <p>• Alibaba Qwen [31]: Qwen 2.5 Instruct (3B).</p>
      <p>All models were used without post-processing, reclassification , or ensemble methods. Fewer than
0.5% of API requests returned empty strings, which were retained as-is in our submissions to preserve
the authenticity of model outputs.</p>
      <sec id="sec-5-1">
        <title>5.2.1. Zero-shot Prompting (Zero-shot Setting)</title>
        <p>For zero-shot inference, we developed language-specific prompts that define the normalization task
without providing examples. The English prompt (Figure 2) served as a template, which was then
translated into other languages to ensure consistent task framing. The complete set of prompts is
available in Appendix A.</p>
        <sec id="sec-5-1-1">
          <title>You have received an informal and disorganized social media post. Summarize this post into a clear and concise statement, without adding any new information.</title>
          <p>Post: {original_post}</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>Normalized statement:</title>
      </sec>
      <sec id="sec-5-3">
        <title>5.2.2. Few-shot Prompting (Monolingual Setting)</title>
        <p>In the few-shot setting (applied to monolingual languages as an alternative to SLM fine-tuning ), we
investigated the impact of varying the number of in-context demonstrations by evaluating prompts with
3, 5, and 10 examples (shots). These examples were selected from the training data of the respective
language using several distinct strategies to assess their influence on model performance :
• Random Selection: Examples were drawn uniformly at random without replacement.
• Mixed Dificulty : Examples combined ’easy’ and ’hard’ posts (as defined below ) to foster
robustness.
• Hard Only: Examples exclusively featured ’hard’ posts (as defined below ) to test model
performance on challenging cases.
• HDBSCAN Top-k Prototypes: Examples comprised prototypes from the  largest HDBSCAN
clusters (e.g.,  = 3 or  = 5 ), aiming for diverse semantic coverage.</p>
        <p>For the dificulty-stratified strategies , example dificulty was determined by calculating the METEOR
score of the original post against its normalized version for every example in the training dataset.
Training examples with the lowest METEOR scores (relative to other examples in the same dataset)
were heuristically classified as ’hard’, hypothesized to be more challenging for the model or to represent
lower-quality reference outputs. Conversely, ’easy’ examples were those with the highest METEOR
scores.</p>
        <p>For HDBSCAN-based strategies, semantically diverse prototypes were selected by first converting
posts into sentence embeddings using the ‘paraphrase-multilingual-MiniLM-L12-v2‘ model, a distilled
Transformer architecture optimized for multilingual sentence-level representations [32, 33]. HDBSCAN
(with min_cluster_size=5) then clustered these embeddings. The post closest to each cluster’s
centroid was designated as a prototype. The ’Top-k Prototypes’ strategy used prototypes from the 
largest clusters, supplemented with random examples if the number of clusters was less than  . This
approach targets broad semantic coverage with minimal curation.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>This section presents the performance of team AKCIT-FN in the CLEF-2025 CheckThat! Task 2 on
Claim Normalization. Our language-adaptive framework achieved podium finishes (top three) in 15
out of the 20 languages evaluated. Specifically , our submissions secured:
• Second place in 8 languages: Tamil, Thai, Punjabi, Telugu, Greek, Romanian, Dutch, and Korean.
• Third place in 7 languages: Portuguese, Spanish, French, Indonesian, Bengali, Polish, and</p>
      <p>German.</p>
      <p>Table 5 provides a detailed breakdown of our best submission for each language, including the
strategy employed, model specifications , average METEOR score, and the oficial ranking assigned by
the organizers.</p>
      <p>Analysis of our submissions reveals distinct trends based on data availability. For the monolingual
setting (Table 5(a)), where in-domain training data was available, fine-tuned Small Language Models
(SLMs) consistently outperformed few-shot Large Language Model (LLM) prompting strategies. This is
evidenced by all our best-evaluated submissions in this setting utilizing SLM fine-tuning .</p>
      <p>Notably, the Varta T5 base model (395M parameters) proved highly efective for Indic languages ,
securing 2nd place for Tamil (0.5197 METEOR) and Punjabi (0.3038 METEOR), and was also the model
of choice for Hindi and Marathi (both achieving 5th place). Furthermore, SLM fine-tuning led to 3rd
place finishes for Portuguese , Spanish, Indonesian, French, Polish, and German, employing various
language-specific T5-based models with parameter counts ranging from 60M (T5 German small) to
approximately 275M (PLT5 Base).</p>
      <p>Conversely, in the zero-shot setting (Table 5(b)), characterized by the absence of language-specific
training data, inference with pre-trained Large Language Models (LLMs) demonstrated strong
generalization capabilities. Our top performances in this category were achieved using models such as
Qwen 2.5 Instruct (3B parameters), which secured 2nd place for Telugu (0.5176 METEOR) and 3rd for
Bengali (0.2916 METEOR), and GPT-4o Mini, which achieved 2nd place for Greek, Romanian, Dutch,
and Korean. These results underscore the utility of LLMs for rapid adaptation to new languages where
specialized training data is scarce.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>This paper detailed AKCIT-FN’s participation in the CLEF-2025 CheckThat! Task 2 on multilingual claim
normalization. Our approach involved a language-adaptive strategy: fine-tuning language-specific or
multilingual Small Language Models (SLMs) for languages with training data, and employing zero-shot
prompting with Large Language Models (LLMs) for languages without such data.</p>
      <p>Monolingual
(training data English
available)
Zero-shot
(no training
data)</p>
      <p>Our submissions demonstrated strong performance, achieving podium finishes (top three) in 15 out
of the 20 languages. Notably, fine-tuned SLMs excelled in supervised settings , while LLMs proved
efective for zero-shot generalization , securing five second-place and one third-place finish among the
seven zero-shot languages. For Portuguese, our primary development language, our best system (Mono
PTT5 base) ranked third with a METEOR score of 0.5290.</p>
      <p>Overall, these results underscore the complementary strengths of SLM fine-tuning when in-domain
data is available and the powerful generalization capabilities of LLMs for rapid deployment in zero-shot
scenarios for complex NLP tasks like claim normalization. A limitation of our work, however, is the
lack of a qualitative error analysis or a discussion of failure cases, which would be a valuable direction
for future investigation. Our code and configurations are publicly available to facilitate further research .</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This work has been fully funded by the project Computational Techniques for Multimodal Data Security
and Privacy supported by Advanced Knowledge Center in Immersive Technologies (AKCIT), with
ifnancial resources from the PPI IoT/Manufatura 4 .0 / PPI HardwareBR of the MCTI grant number
057/2023, signed with EMBRAPII.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Gemini 2.5 Pro [27]
(gemini-2.5-pro-exp-0325), Claude Sonet 4 [34] (claude-sonnet-4-20250514), GPT-4o [28] (gpt-4o-2024-05-13) in order
to: Paraphrase and reword, Improve writing style, Abstract drafting , and Peer review simulation.
After using this tool/service , the author(s) reviewed and edited the content as needed and take(s) full
responsibility for the publication’ s content.
Verlag, Berlin, Heidelberg, 2020, p. 215–236. URL: https://doi.org/10.1007/978-3-030-58219-7_17.
doi:10.1007/978- 3- 030- 58219- 7_17.
[9] P. Nakov, G. Da San Martino, T. Elsayed, A. Barrón-Cedeño, R. Míguez, S. Shaar, F. Alam,
F. Haouari, M. Hasanain, W. Mansour, B. Hamdan, Z. S. Ali, N. Babulkov, A. Nikolov, G. K.
Shahi, J. M. Struß, T. Mandl, M. Kutlu, Y. S. Kartal, Overview of the clef–2021 checkthat! lab
on detecting check-worthy claims, previously fact-checked claims, and fake news, in: K. S.
Candan, B. Ionescu, L. Goeuriot, B. Larsen, H. Müller, A. Joly, M. Maistro, F. Piroi, G. Faggioli,
N. Ferro (Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction, Springer
International Publishing, Cham, 2021, pp. 264–291.
[10] D. S. Nielsen, R. McConville, Mumin: A large-scale multilingual multimodal fact-checked
misinformation social network dataset, in: Proceedings of the 45th International ACM SIGIR Conference
on Research and Development in Information Retrieval, SIGIR ’22, Association for Computing
Machinery, New York, NY, USA, 2022, p. 3141–3153. URL: https://doi.org/10.1145/3477495.3531744.
doi:10.1145/3477495.3531744.
[11] I. Singh, C. Scarton, X. Song, K. Bontcheva, Breaking language barriers with mmtweets: Advancing
cross-lingual debunked narrative retrieval for fact-checking, 2024. URL: https://arxiv.org/abs/2308.
05680. arXiv:2308.05680.
[12] S. Banerjee, A. Lavie, METEOR: An automatic metric for MT evaluation with improved correlation
with human judgments, in: J. Goldstein, A. Lavie, C.-Y. Lin, C. Voss (Eds.), Proceedings of the
ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or
Summarization, Association for Computational Linguistics, Ann Arbor, Michigan, 2005, pp. 65–72.</p>
      <p>URL: https://aclanthology.org/W05-0909/.
[13] D. Carmo, M. Piau, I. Campiotti, R. Nogueira, R. Lotufo, Ptt5: Pretraining and
validating the t5 model on brazilian portuguese data, 2020. URL: https://arxiv.org/abs/2008.09144.
arXiv:2008.09144.
[14] H. W. Chung, L. Hou, S. Longpre, B. Zoph, Y. Tai, W. Fedus, Y. Li, X. Wang, M. Dehghani,
S. Brahma, A. Webson, S. S. Gu, Z. Dai, M. Suzgun, X. Chen, A. Chowdhery, A. Castro-Ros,
M. Pellat, K. Robinson, D. Valter, S. Narang, G. Mishra, A. Yu, V. Zhao, Y. Huang, A. Dai, H. Yu,
S. Petrov, E. H. Chi, J. Dean, J. Devlin, A. Roberts, D. Zhou, Q. V. Le, J. Wei, Scaling
instructionifnetuned language models , J. Mach. Learn. Res. 25 (2024).
[15] M. Piau, R. Lotufo, R. Nogueira, ptt5-v2: A closer look at continued pretraining of t5 models for
the portuguese language, in: A. Paes, F. A. N. Verri (Eds.), Intelligent Systems, Springer Nature
Switzerland, Cham, 2025, pp. 324–338.
[16] Adalberto Ferreira Barbosa Junior, bart-base-portuguese (revision 149de72), 2024. URL: https:
//huggingface.co/adalbertojunior/bart-base-portuguese. doi:10.57967/hf/3264.
[17] E. M. B. Nagoudi, A. Elmadany, M. Abdul-Mageed, AraT5: Text-to-text transformers for Arabic
language generation, in: S. Muresan, P. Nakov, A. Villavicencio (Eds.), Proceedings of the
60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers),
Association for Computational Linguistics, Dublin, Ireland, 2022, pp. 628–647. URL: https://
aclanthology.org/2022.acl-long.47/. doi:10.18653/v1/2022.acl- long.47.
[18] guillaumephd, T5-french-base model: A t5 model trained on french data only, 2024. URL: https:
//huggingface.co/guillaumephd/t5-french-base.
[19] Shahm, t5-seven-epoch-base-german, 2023. URL: https://huggingface.co/Shahm/t5-small-german.
[20] R. Aralikatte, Z. Cheng, S. Doddapaneni, J. C. K. Cheung, Varta: A large-scale headline-generation
dataset for Indic languages, in: A. Rogers, J. Boyd-Graber, N. Okazaki (Eds.), Findings of the
Association for Computational Linguistics: ACL 2023, Association for Computational
Linguistics, Toronto, Canada, 2023, pp. 3468–3492. URL: https://aclanthology.org/2023.findings-acl.215/.
doi:10.18653/v1/2023.findings- acl.215.
[21] C. Wirawan, Indonesian T5 Summarization Base Model, https://huggingface.co/cahya/
t5-base-indonesian-summarization-cased, 2021. Accessed: 2025-05-29.
[22] A. Chrabrowa, Ł. Dragan, K. Grzegorczyk, D. Kajtoch, M. Koszowski, R. Mroczkowski, P. Rybak,
Evaluation of transfer learning for Polish with a text-to-text model, in: N. Calzolari, F. Béchet,
[32] W. Wang, H. Bao, S. Huang, L. Dong, F. Wei, MiniLMv2: Multi-head self-attention relation
distillation for compressing pretrained transformers, in: C. Zong, F. Xia, W. Li, R. Navigli
(Eds.), Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Association
for Computational Linguistics, Online, 2021, pp. 2140–2151. URL: https://aclanthology.org/2021.
findings-acl.188/. doi:10.18653/v1/2021.findings- acl.188.
[33] N. Reimers, I. Gurevych, Sentence-BERT: Sentence embeddings using Siamese BERT-networks, in:
K. Inui, J. Jiang, V. Ng, X. Wan (Eds.), Proceedings of the 2019 Conference on Empirical Methods
in Natural Language Processing and the 9th International Joint Conference on Natural Language
Processing (EMNLP-IJCNLP), Association for Computational Linguistics, Hong Kong, China, 2019,
pp. 3982–3992. URL: https://aclanthology.org/D19-1410/. doi:10.18653/v1/D19- 1410.
[34] Anthropic, System Card: Claude Opus 4 &amp; Claude Sonnet 4, Technical Report, Anthropic, 2025.</p>
      <p>URL: https://www-cdn.anthropic.com/4263b940cabb546aa0e3283f35b686f4f3b2ff47.pdf.</p>
    </sec>
    <sec id="sec-10">
      <title>A. Zero-shot prompts</title>
      <p>This appendix lists the prompts used for the zero-shot claim normalization task. The English prompt in
Figure 2 served as the template and was translated into the seven target languages. The {post_text}
placeholder is replaced with the social media post text during inference.</p>
      <p>• Czech</p>
      <p>Dostanete neformální a neuspořádaný příspěvek ze sociálních sítí. Shrňte jej do jasného a
stručného tvrzení, bez přidávání dalších informací.</p>
      <p>Příspěvek: {post_text}
• Greek
Σου δίνεται μια ανοργάνωτη και ανεπίσημη ανάρτηση στα κοινωνικά δίκτυα. Περίληψέ την
σε μια σαφή και συνοπτική δήλωση, χωρίς να προσθέσεις επιπλέον πληροφορίες.
Ανάρτηση: {post_text}
• Dutch</p>
      <p>Je ontvangt een informeel en ongeorganiseerd bericht op een sociaal netwerk. Vat het samen in
een duidelijke en beknopte verklaring zonder extra informatie toe te voegen.</p>
      <p>berichten:{post_text}
• Korean
비정형적이고 비공식적인 소셜 미디어 게시물이 주어집니다. 이를 명확하고 간결한 주장
으로 요약하십시오. 추가 정보는 포함하지 마십시오.</p>
      <p>게시물:{post_text}
• Romanian</p>
      <p>Prompt: Primești o postare informală și dezorganizată de pe o rețea socială. Rezum-o într-o
afirmație clară și concisă , fără a adăuga informații.</p>
      <p>Postare: {post_text}
• Tegulu
మీకు ఒక అసంఘటితమైన, అనౌపచారికమైన సోషల్ మీడియాలో పోస్ట్ ఇవ్వబడుతుంది.
స్పష్టమైన మరియు సంక్షిప్తమైన ప్రకటనగా మార్చండి, అదనపు సమాచారం ఇవ్వకుండా.
పోస్ట్:{post_text}
దీనిని
আপনি একটি অগোছালো এবং অনানুষ্ঠানিক সোশ্যাল মিডিয়া পোস্ট পাচ্ছেন। এটি একটি
স্পষ্ট এবং সংক্ষিপ্ত দাবিতে রূপান্তর করুন, কোনো অতিরিক্ত তথ্য ছাড়াই। পোস্ট:
{post_text}</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>From chaos to clarity: Claim normalization to empower fact-checking, in: Findings of the Association for Computational Linguistics: EMNLP 2023, Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>6594</fpage>
          -
          <lpage>6609</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Venktesh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! Lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: J.
          <string-name>
            <surname>Carrillo-de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. V.</surname>
          </string-name>
          ,
          <article-title>The clef-2025 checkthat! lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: C.
          <string-name>
            <surname>Hauf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Kazai</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nardini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pinelli</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Silvestri</surname>
          </string-name>
          , N. Tonellotto (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2025</year>
          , pp.
          <fpage>467</fpage>
          -
          <lpage>478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! lab task 2 on claim normalization</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.), Working Notes of CLEF 2025 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2025</year>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Konstantinovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Price</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Babakar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zubiaga</surname>
          </string-name>
          , Toward automated factchecking:
          <article-title>Developing an annotation schema and benchmark for consistent automated claim detection</article-title>
          ,
          <source>Digital Threats</source>
          <volume>2</volume>
          (
          <year>2021</year>
          ). URL: https://doi.org/10.1145/3412869. doi:
          <volume>10</volume>
          .1145/3412869.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Màrquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Atanasova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kyuchukov</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Da San Martino, Overview of the clef-2018 checkthat! lab on automatic identification and verification of political claims</article-title>
          , in: P. Bellot,
          <string-name>
            <given-names>C.</given-names>
            <surname>Trabelsi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mothe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Murtagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Soulier</surname>
          </string-name>
          , E. SanJuan, L. Cappellato, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction</source>
          , Springer International Publishing, Cham,
          <year>2018</year>
          , pp.
          <fpage>372</fpage>
          -
          <lpage>387</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          , G. Da San Martino, P. Atanasova,
          <article-title>Overview of the clef-2019 checkthat! lab: Automatic identification and verification of claims , in: Experimental IR Meets Multilinguality</article-title>
          , Multimodality, and
          <source>Interaction: 10th International Conference of the CLEF Association, CLEF</source>
          <year>2019</year>
          , Lugano, Switzerland, September 9-
          <issue>12</issue>
          ,
          <year>2019</year>
          , Proceedings, Springer-Verlag, Berlin, Heidelberg,
          <year>2019</year>
          , p.
          <fpage>301</fpage>
          -
          <lpage>321</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -28577-7_
          <fpage>25</fpage>
          . doi:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>030</fpage>
          - 28577- 7_
          <fpage>25</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          , G. Da San Martino, M. Hasanain,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hamdan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. S.</given-names>
            <surname>Ali</surname>
          </string-name>
          , Overview of checkthat! 2020:
          <article-title>Automatic identification and verification of claims in social media</article-title>
          ,
          <source>in: Experimental IR Meets Multilinguality, Multimodality, and Interaction: 11th International Conference of the CLEF Association, CLEF</source>
          <year>2020</year>
          , Thessaloniki, Greece,
          <source>September 22-25</source>
          ,
          <year>2020</year>
          , Proceedings, Springer-
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>