<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the CLEF-2024 CheckThat! Lab Task 6 on Robustness of Credibility Assessment with Adversarial Examples (InCrediblAE)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Piotr Przybyła</string-name>
          <email>piotr.przybyla@upf.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ben Wu</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexander Shvets</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yida Mu</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kim Cheng Sheang</string-name>
          <email>kimcheng.sheang@upf.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xingyi Song</string-name>
          <email>x.song@shefield.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Horacio Saggion</string-name>
          <email>horacio.saggion@upf.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Computer Science, Polish Academy of Sciences</institution>
          ,
          <addr-line>Warsaw</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universitat Pompeu Fabra</institution>
          ,
          <addr-line>Barcelona</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Shefield</institution>
          ,
          <addr-line>Shefield</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Task 6 at CheckThat! Lab, organised at CLEF-2024, is devoted to assessing the robustness of misinformation detection solutions implemented as text classification models. The participants of the task were provided with prediction models and data examples for several problems of credibility estimation and their goal was to come up with adversarial examples (AEs): small modifications to the provided text fragments, such that the original meaning is preserved, but the victim classifier changes its decision. The evaluation involved five domains (detection of: biased news, propaganda techniques, false claims, rumours and COVID-19 misinformation) and three classifiers (BiLSTM, BERT and adversarially fine-tuned RoBERTa). Six teams participated in the task, representing a variety of approaches and substantially outperforming previous AE generation solutions. We also performed manual evaluation, which highlighted some modification techniques that are particularly likely to pass unnoticed by human readers. Overall, the task results emphasise the need to assess the robustness of text classification solutions before implementing them in content filtering on large platforms, such as social media.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;adversarial examples</kwd>
        <kwd>robustness</kwd>
        <kwd>misinformation detection</kwd>
        <kwd>credibility</kwd>
        <kwd>text classification</kwd>
        <kwd>natural language processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The challenges of misinformation have been taken up with great energy and vigour by the NLP and IR
communities. The main reasons for such enthusiastic adoption of the new tasks are wide availability of
textual data to train on and tantalisingly simple dichotomy of fake vs. real, clearly fitting the familiar
task of binary classification. Among a great deal of work in the domain [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], this framework has also
enabled numerous shared tasks, including detecting hyperpartisan news [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], propaganda [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], bots [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
false claims [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and more. These research results quickly found applications in content moderation for
large media platforms, which increasingly rely on ML tools to support, but also to replace the human
efort [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        However, a shared task framework is far from the real-world application scenario, where the test
data are not fixed, but are generated continuously by users. This means that if a malicious actor sees
their non-credible content rejected by the system, they are likely to try to modify it to pass the filters,
rather than simply abandon their goals. Unfortunately, the deep learning architectures, which many of
the best-performing solutions use, are known for their susceptibility to adversarial examples, i.e. data
instances modified with the intent of fooling a classifier [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. While discovering adversarial examples
for text is more challenging than in other domains, it is definitely possible [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Thus, investigating
the robustness of text credibility assessment solutions is indispensable for making them applicable in
real-world adversarial scenarios.
      </p>
      <p>
        Here we report on the shared task on Investigating Robustness of Credibility Assessment with
Adversarial Examples (InCrediblAE), which was organised as Task 6 of the CheckThat! 2024 evaluation
lab [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ] at the CLEF 2024 conference. In InCrediblAE, participants get access to the following
resources for each domain:
1. three victim classifiers , assessing the credibility of the input text and returning a score,
2. an attack dataset, including around 400 instances unseen by the classifier in training.
In the task we take into account five domains, corresponding to important challenges in the credibility
assessment – see section 2 for details.
      </p>
      <p>The goal of the participants is to make modification to the instances in the attack dataset, turning
them into adversarial examples. Each adversarial example (AE) is evaluated on meaning preservation, i.e.
how similar it is to the original; and classifier confusion , i.e. whether the output of the victim classifier
is diferent than for the original.</p>
      <p>
        The evaluation consists of two stages. The automatic evaluation follows the framework established
in the field of adversarial learning, with the above factors assessed through automatic measures, i.e.
BODEGA score [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], leading to a ranking list – see section 3.1. In the manual evaluation we use human
judgement to assess the semantic similarity between attack sentences and their original counterparts.
This process aims to highlight potential errors arising from automatic evaluation metrics, as well as to
create a high-quality text similarity dataset for the future development and evaluation of metrics – see
section 3.2.
      </p>
      <p>The task has attracted six teams submitting various approaches (section 4.1), some of which have
clearly outperformed previous solutions applied to the same problems (section 4.2), generally confirming
the high vulnerability of popular text classifiers to adversarial attacks. What is more, the results of the
manual evaluation (section 4.3) highlight cases where AEs might appear far from the original text if
judged by automatic measures, but are in fact quite convincing to human annotators. The results of the
manual annotation are made openly available for future research1. We also share the code2 and data3
for the automatic evaluation.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task data</title>
      <p>
        The shared task uses the foundation of the BODEGA framework [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], which has been created to enable
systematic robustness testing in the area of misinformation detection. Within this framework, several
domains are available, each organised around a credibility assessment problem, defined as a binary text
classification task. Within a domain, an expert-annotated corpus of documents is used to train victim
classifiers (train subset) and test the attack performance (attack subset).
      </p>
      <p>
        The participants are provided with Python code, hosted on Google Colab, allowing to interact with
the victim models in an attack scenario through OpenAttack interface [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Their goal is to prepare
a procedure that modifies the text fragments in the attack dataset to achieve a diferent decision of a
classifier with a minimal meaning alteration. The participant submission includes the AEs, as well as
the number of victim model queries needed to find them.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Datasets</title>
        <p>Four of the domains in InCrediblAE have been prepared in BODEGA based on previously published
corpora. The final one (C19) is new and was not available before.</p>
        <p>Style-based news bias assessment (HN) is a task of verifying credibility of a news article based
on its overall writing style. It relies on previous work indicating that stylistic analysis of fake news</p>
        <sec id="sec-2-1-1">
          <title>1https://github.com/GateNLP/CLEF2024_InCrediblAE_Manual_Evaluation_Dataset 2https://github.com/piotrmp/BODEGA 3https://gitlab.com/checkthat_lab/clef2024-checkthat-lab/-/tree/main/task6</title>
          <p>
            outlets can be used to distinguish them from credible sources [
            <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
            ]. The corpus contains news
bias annotations assigned at the level of the source (whole website) by journalists from BuzzFeed and
MediaBiasFactCheck.com4. For the purpose of BODEGA, 10% of the training instances were used and
the non-credibility label was assigned to articles from sources marked as hyperpartisan, both left- and
right-wing.
          </p>
          <p>
            Propaganda detection (PR) is focused on recognising specific manipulation techniques appealing
to emotions [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ], for example name-calling, flag-waving or straw-man fallacy. This approach has
the advantage of being fine-grained by highlighting manipulative fragments in text, akin to the NER
(Named Entity Recognition) tasks. We rely on the token-level annotations of 14 techniques, marked
by professional annotators for the SemEval 2020 Task 11 (Detection of Propaganda Techniques in News
Articles) [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ], for which the training set is public5. In order to cast the task as binary text classification for
BODEGA, the corpus was split into sentences and those including some tokens marked as propaganda
were labelled as non-credible.
          </p>
          <p>
            Fact checking (FC) is an approach to misinformation detection based on extracting claims made in
a piece of text and verifying them with respect to a trusted knowledge base [
            <xref ref-type="bibr" rid="ref17">17</xref>
            ]. In order to represent
the problem as binary classification, we focus on the final stage of the workflow, when a claim is
compared to relevant evidence from the knowledge base, which either confirms its validity or refutes
it. In BODEGA, the data from the FEVER shared task [
            <xref ref-type="bibr" rid="ref18">18</xref>
            ] is used, consisting of claims that were
paired with relevant passages from Wikipedia articles. The instances where a claim is supported by the
evidence were labelled as credible, and those when it is refuted as non-credible.
          </p>
          <p>
            Rumour detection (RD) is aimed at detecting information spreading widely over social media
despite it not coming from a credible source. Rumours can be detected using many indicators [
            <xref ref-type="bibr" rid="ref19">19</xref>
            ],
but here we focus on the textual content of a social media post, as well as the reactions of other
users. In BODEGA, this is achieved thanks to the augmented dataset of rumours and non-rumours for
rumour detection [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ], created from Twitter threads relevant to six real-world events, labelled by experts
according to the source reliability of the initial post. One of the events (Charlie Hebdo shooting) was
set aside as the attack dataset.
          </p>
          <p>
            COVID-19 misinformation detection (C19) focuses on binary classification 6 of misinformation
related to COVID-19 [
            <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
            ]. Given a known false claim about the disease, the task is to determine
whether a user’s tweet supports that false claim. If so, the tweet is classified as COVID-19 misinformation
(positive class). Alternative responses, such as contradicting, questioning, commenting, or irrelevance
regarding the false claim are reserved for the negative class. Appendix B provides examples from the
dataset.
          </p>
          <p>Table 1 summarises the information on datasets, including the sizes of subsets: training (for training
victim classifiers), development (reserved for future use) and attack (to be modified into AEs), as well as
the percentage of positive (non-credible) instances.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>4https://zenodo.org/record/1489920 5https://zenodo.org/record/3952415 6We merged IRRELEVANT and DEBUNK class as non-misinfo to covert original dataset into a binary classification</title>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Victim classifiers</title>
        <p>Training datasets were used to prepare victim models, representing popular approaches to text
classification. Two of the models (BiLSTM and BERT) were trained as in BODEGA framework, but the surprise
classifier was trained specifically for the shared task and it was revealed to the participants in the test
phase, one week before the submissions.</p>
        <p>
          BiLSTM classifier consists of an embedding layer (token representations of size 32), two LSTM [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]
layers (forwards and backwards, hidden representation of size 128) and a dense linear layer converting
the text fragment representation (of size 256) into two-class probability, normalised using softmax.
        </p>
        <p>
          BERT classifier is a bert-base-uncased model [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] from the HuggingFace Transformers library
[
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], fine-tuned for sequence classification using Adam optimiser with linear weight decay [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] for 5
epochs.
        </p>
        <p>
          Surprise classifier is a RoBERTa model [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ], i.e. roberta-base from HuggingFace Transformers,
adversarially-trained to be more robust to adversarial attacks. We use data augmentation to improve
robustness: First, the model is fine-tuned for one epoch on the train dataset, then adversarial examples
are generated from the entire train dataset using BERT-ATTACK [28], and then the model is fine-tuned
for one epoch on a combination of the train dataset and the successful adversarial examples. We train
with constant learning rate 2 × 10− 5 and the Adam optimiser. We use batch of size 32 for all tasks
except PR, which uses a batch of size 64. Due to computational constraints, for HN we only generate
adversarial examples from a subset (6000 samples) of the training data.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Evaluation</title>
      <p>The evaluation procedure consists of two stages. Firstly, the BODEGA framework is used to automatically
assess the attack efectiveness of each participant in 15 scenarios (5 domains × 3 victims). The average
BODEGA score is used to create the leaderboard, expressing the overall performance. Secondly, the
task most challenging for automatic evaluation (fact-checking) is used to perform manual annotation of
meaning preservation in selected instances.</p>
      <sec id="sec-3-1">
        <title>3.1. Automatic evaluation</title>
        <p>
          In automatic evaluation, when an example  is modified into AE * , the quality of the transformation
is assessed through BODEGA score defined as follows [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]:
        </p>
        <p>BODEGA_score(, * ) = Con_score(, * ) × Sem_score(, * ) × Char_score(, * ),
where:
• Con_score, i.e. confusion score, takes value of 1 when the attacked classifier predicts a diferent
class for * than it did for , and 0 otherwise.
• Sem_score, i.e. semantic similarity score, is a measure of meaning preservation between  and
* , computed using the BLEURT [29] evaluation measure (BLEURT-20 variant), clipped to the
(0-1) range.
• Char_score, i.e. character similarity score, is a measure of similarity of  and * as character
sequences, computed through Levenshtein distance [30], scaled to (0-1) similarity.</p>
        <p>We can see that an AE will be ranked highly if it changes the output of the classifier
(Con_score(, * ) = 1), but at the same time preserves both the meaning (Sem_score(, * ) ≈ 1)
and the appearance (Char_score(, * ) ≈ 1) of the original text.</p>
        <p>To measure the overall attack success in a particular scenario, the BODEGA score averaged over all
instances in the attack set is employed. However, the constituent scores, also averaged over the dataset,
can be used to understand the results. We also report the number of queries that is needed (on average)
for a single AE to be found.</p>
        <p>Definitions
Extremely unconfident about the annotation (I’m really unsure about the annotation. It may belong
to another category as well, you may wish to discard this instance from the training.)
Not confident about the annotation (I’m not sure about the annotation, it seems it also belongs to
other categories, but you can still include this instance as a “silver standard instance” in training.)
Pretty confident about the annotation (I’m pretty sure about the annotation, but might be in high
chance other annotators may label it in a diferent category.)
Fairly confident about the annotation (I’m confident about the annotation, but might be in small
chance other annotators may label it in a diferent category.)</p>
        <p>Extremely confident about the annotation (I’m certain about the annotation without a doubt.)</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Manual evaluation</title>
        <p>The goal of the manual evaluation is to highlight the cases where the automatic evaluation measures,
especially regarding semantic similarity, might not be an accurate representation of human reception of
the adversarially modified content. To that end, we have selected the samples from the fact-checking
domain and the surprise victim, where even small changes in text can alter the meaning and, consequently,
the credibility label.</p>
        <p>Task Description We aim to gather assessments regarding the semantic similarities between attack
samples and the original samples. Participants in the shared task are requested to dedicate approximately
60 minutes to this manual evaluation (i.e., 100 samples per participant), which are conducted using an
open-source, collaborative annotation platform, i.e., GATE Teamware 2 [32]. Judges rate the sample
pairs based on the following scale: (a). Preserve the Semantic Meaning, (b). Change the Semantic
Meaning, and (c). No sense. Table 2 demonstrates categories and descriptions. Similar to the work of
Mu et al. [31], annotators are required to indicate the confidence level (see Table 3) of their assigned
class in the ‘Confidence Row’.</p>
        <p>Data Sampling We randomly select 100 paired samples (i.e., original and modified texts) from
each submission, resulting in a total of 600 paired samples. Note that only the successful attacks are
considered.</p>
        <p>Annotator Training We train the annotators by providing a training document detailing the
annotation pipeline, which includes (i) a step-by-step tutorial for using the GATE Teamware platform, (ii)
a user information sheet to inform about any potential issues and risks that may occur during data
annotation, and (iii) a user consent sheet as required by the ethical approval from the University of
Shefield, where the annotation was performed.</p>
        <p>A total of 12 annotators (i.e., 6 participants and 6 organisers) were recruited to manually annotate the
paired samples. These 12 annotators were further divided into 6 separate groups (i.e., two annotators
per group). In each group, 100 tweets were assigned to each annotator. Finally, this process yielded 100
double-annotated paired samples from each group, resulting in 600 double-annotated samples in total.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Annotation Methodology and Quality Assurance All samples are double-annotated by the shared</title>
        <p>task organizers and participants. Briefly, each paired sample is annotated by one participant and one
shared task organiser. A third annotator from the shared-task organisers is used to resolve any conflicts.
Given that there are three categories in total, the annotation with the highest confidence score will be
considered in the case of three difering annotations.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>Here we outline the results of the InCrediblAE shared task in three steps: first, we describe the solutions
submitted by the participants (section 4.1), then we present the results of the automatic (section 4.2)
and manual evaluation (section 4.3).</p>
      <sec id="sec-4-1">
        <title>4.1. Participating solutions</title>
        <p>The SINAI team [33] proposes a method for adversarial attacks based on the substitution of characters
by homoglyphs (e.g. characters which resemble the target such as l ≈ 1). The method uses exhaustive
search with two variants: with memory and without memory. They ground their approach in the fact
that homoglyphs could deceive the human eye while at the same time provoke a Large Language Model
classifier to reverse its prediction due to the presence of an unexpected token. According to the oficial
leader board, the approach is sub-optimal, obtaining the last rank in the task according to the oficial
evaluation metrics. However, human evaluation of content preservation is ranked high.</p>
        <p>The MMU_NLP participation [34] features a system to attack classifiers based on lexical substitution
and character replacement. The proposed method searches for candidate words to attack followed
by a word replacement mechanism. The word search mechanism masks words to check their
vulnerability with those words having a high impact on the classifier performance retained for the attack.
The replacement step uses homoglyphs for character replacement or lexical substitution. Character
replacement is tested in two diferent conditions: random character attack or begin/end of word attack.
The lexical replacement attack uses a large language model to retrieve a word similar to the target
word. Overall result fall short compared to other participants, however the proposed methods improves
over the baseline in several settings. The character attack method seems more efective than the word
replacement approach.</p>
        <p>The Palöri team [35] proposes an approach to identify vulnerable words by computing (relying on a
masked language model) the diference between the probability distribution of the original sentence
and the sentence with a word masked. These diferences are used as scores to rank words by their
"vulnerability" according to the model. The ranked words (most to least vulnerable) are then replaced
using a word from a list of substitutes proposed by the language model. The sentence with the replaced
word is used to attack the victim classifier. In case of success, the new sentence is returned, otherwise
the method loops using the sentence with the "best" possible substitution (i.e. one which reduced the
victim’s confidence the most). The method produces successful attacks which however do not preserve
the original sentence’s meaning. To address this problem a "synonym" dictionary is created, using
GloVe embeddings and the aclImbd dataset, to draw substitutes from. The new method only contributes
minor improvements over the masked language model.</p>
        <p>
          The solution of the OpenFact team [36] consists in a coupling of various word-substitution approaches
in an ensemble in such a way that if the first approach does not succeed in changing the classifier’s
decision, then the second one is called. In particular, a modification of BERT-ATTACK [ 28] was
proposed (it features a change in parameter values, an alternative selection of a replacement position
by an exhaustive search of candidates that provide the largest diference in probabilities for a predicted
class, and an iterative replacement from 1 to 7 words, including punctuation and digits at the latter,
until the success of an attack) and backed up by a genetic algorithm [37] realised in the OpenAttack
framework [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Another ensemble was compiled of a proposed greedy search by word swap with
synonyms in the word embedding space (prebuilt "counter-fitted" GloVe embeddings [ 38]) and another
model available in the OpenAttack, TextFooler [39], which unlike other similar approaches replaces
words in agreement with the syntax of the attacked text. Apart from ensemble models, approaches
from the TextAttack framework [40] were used. They demonstrated superior performance over the
baseline methods in automatic scores. In particular, CLARE [41] – a model that implements a special
mask-then-infill procedure that incorporates replace/insert/merge operations allowing for outcomes
varied in length – was applied for PR, FC and C19 tasks and consistently yielded better results in all
automatic scores. Overall, this solution gained the best automatic scores in most of the domains for
most of the victims, which made it the first in the leaderboard created by averaging the scores across
all scenarios. However, it was ranked very low within the human evaluation, as in the majority of the
cases the meaning of the text was changed.
        </p>
        <p>
          The TurQUaz [
          <xref ref-type="bibr" rid="ref28">42</xref>
          ] team leverages a genetic algorithm to look for a combination of character
modifications. The modifications that are introduced using a mutation operator include homoglyph replacement,
three options of word splitting (random, favouring existing words in the subword outcome, and special
heuristic-wise), insertion or removal of individual random letters, and shufling the order of the letters
within the word. The search is carried out until the first flip is found. Only the use of homoglyphs and
word splits proved to be eficient. Apart from the genetic algorithm, the team experimented with attacks
made by utilizing large language models (LLMs) such as Llama 37 and Mistral8. Three approaches have
been tried: (i) prompting a model for text paraphrasing, (ii) leveraging a model for identifying words to
be changed, and (iii) generating adversarial examples with one LLM and verifying whether an attack is
going to be successful with another LLM. None of the approaches outperformed the genetic algorithm,
however, the team believes in the potential of LLMs and suggests fine-tuning the models specifically for
this task to improve their performance in the future. The primary solution has been ranked third based
on the BODEGA score.
        </p>
        <p>
          The TextTrojaners team [
          <xref ref-type="bibr" rid="ref29">43</xref>
          ] introduces a BeamAttack method that makes attacks at a word level using
RoBERTa [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] and a beam search as a backbone to produce contextually appropriate word substitutions.
Beam search algorithm adapted by enabling operations of replacing, skipping, or removing words
allows for generating and evaluating multiple alternative word replacement combinations in a single
run. For the identification of the most vulnerable words to be replaced, the team experiments with
two ranking approaches that use the explainable AI framework LIME [
          <xref ref-type="bibr" rid="ref30">44</xref>
          ] and logit-based importance
scores, as proposed in [28]. In a series of ablation studies, they show that the choice of a ranking method
varies across victims and depends on the dataset. The solution achieved the second-best result on the
BODEGA evaluation metric but gained rather low manual evaluation scores.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Automatic evaluation</title>
        <p>Table 4 includes the results of the evaluation against the BiLSTM victim. Generally, we can see that
diferent approaches dominate in diferent scenarios. However, in every domain, the best BODEGA
score (in boldface) is achieved by a solution submitted to InCrediblAE, rather than a reference solution
from previous work (BERT-ATTACK or DeepWordBug). HN appears to be the easiest domain, with the
leading solution (TextTrojaners) achieving BODEGA score of 0.91 through 100% confusion with 91%
semantic similarity and 99% character similarity. This result, closely followed by OpenFact, is especially
impressive when compared to the scores of BERT-ATTACK (BODEGA score of 0.64). The TextTrojaners
approach also leads in PR, but we need to note its high amount of queries needed – in this case, 593
compared to TurQUaz achieving almost the same result (0.68 instead of 0.70) with six times less queries.
The C19 appears to be the domain most challenging for attacks, although even here we note a vast
improvement over the reference method (OpenFact: 0.72 vs BERT-ATTACK: 0.50).</p>
        <p>In the attacks against the BERT victim, evaluated in table 5, the OpenFact method dominates, ceding
only in FC to homoglyph-based SINAI. We can also note that the best BODEGA score is either equivalent
(for FC, HN and C19) or significantly lower (for PR and RD) than for BiLSTM, indicating higher dificulty
of attacking a Transformer-based classifier. This also results in a higher number of queries necessary to</p>
        <sec id="sec-4-2-1">
          <title>7https://huggingface.co/docs/transformers/model_doc/llama3 8https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2</title>
          <p>FC
RD
HN
ifnd an AE, e.g. while the leading method for HN domain needed 937 queries to obtain the score of 0.91
with BiLSTM, attacking BERT with the same approach requires 4328 queries on average.</p>
          <p>The results for the surprise victim (table 6) show that the adversarially-trained classifier indeed is
more challenging to attack, leading to lower scores in PR and RD domains. However, the general view
remains similar, with OpenFact again dominating with the exception of FC. Their approach clearly works
very well with Transformer-based victims, just with an exception of this single task. The increasing
level of dificulty is again reflected in a rising numbers of queries, reaching a record value of over 15000
for TextTrojaners attacking in the RD domain.</p>
          <p>Finally, table 7 shows the final leaderboard obtained by averaging the BODEGA scores across
FC
RD
HN
victims and domains. We can see that all solutions submitted to the task have beaten the DeepWordBug
reference and most have also outperformed BERT-ATTACK, which is a strong reference point. OpenFact
and TextTrojaners are clear leaders, with the former slightly better-performing, especially against
Transformers victim. However, we need to emphasise that the averaged ranking does not show the
whole picture and various methods work best for various scenarios. For example, SINAI is the best
approach for the BERT-FC combination.</p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Manual Evaluation</title>
        <p>We randomly selected 100 successful adversarial samples from each team’s submission to the scenario
including fact checking and surprise victim. During the evaluation, both the original and adversarial
samples were presented to the annotators, with diferences highlighted. The annotators were asked to
categorise each sample pair into one of the three categories described in Section 3.2. They also provided
a confidence score for each annotation (5: very confident, 1: not confident). Each sample pair was
judged by at least two annotators. The annotation agreement of the initial annotators was 0.52 (Cohen’s
Kappa). A third annotator was invited if there was a conflict between the two initial annotators.</p>
        <p>We determined the ranking of participants by the number of samples that fell into the ’Preserve the
Semantic Meaning’ category, adhering to the principle that a higher count indicates better performance.
This ranking method was used because the task demands that the adversarial samples maintain their
original meaning. Table 8 presents the final manual evaluation results for all participants.</p>
        <p>In general, we observe we observe a discrepancy between the manual and automatic evaluations
(see Table 7). This may be because the manual evaluation task is a fact-checking task. Even a slight
replacement of named entities (such as changing the year from 1990 to 1991) could result in a change of
meaning.</p>
        <p>The leading team (SINAI) in manual evaluation uses an adversarial attack method mainly based on
the substitution of characters with homoglyphs and achieves a 99% score. This suggests that the use of
homoglyphs can successfully deceive the annotator’s eye. Similarly, team MMU employs similar attack
approaches, such as lexical substitution and character replacement, which also achieve a high manual
evaluation score (96%).</p>
        <p>Team TurQUaz proposes a method of inserting white space to split English words, which achieved
third place on the leaderboard (62%). However, this method may sometimes change the meaning of
the original fact-checking document, resulting in a lower manual evaluation score compared to the
methods proposed by teams SINAI and MMU. Besides, by inserting white space in the original text,
the modified text may become non-interpretable by humans, resulting in a high proportion of third
category submissions by Team TurQUaz, i.e., “The sentence does not make any sense.”</p>
        <p>We observe that OpenFact (11%) and TextTrojaners (7%) obtain lower manual evaluation scores. By
manually investigating text modified by Team OpenFact, we notice that some key information, such as
time and location, has been changed. Note that such named entities play a vital role in the context of
fact-checking downstream tasks. Therefore, changes to these named entities result in a higher number
of samples labelled as “Change the Semantic Meaning.” As for the solution of TextTrojaners, the choice
of alternative words solely depends on the context rather than the word to be replaced. This can
significantly deviate the meaning. In addition, the operation of word removal without further word
agreement adjustment may lead to nonsensical sentences.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The first conclusion from the results obtained is clear: the state of the art in AE generation for
misinformation detection, established by previous solutions, has advanced considerably. Various solutions were
submitted to the shared task, but they are based on the established lines of research in the area: word
replacements (preserving meaning similarity) or character replacements (preserving visual similarity).</p>
      <p>The word-level solutions (esp. TextTrojaners and OpenFact) were performing the best in most
scenarios, but not all of them. Fact checking is a clear outlier due to its nature: every word matters,
making it hard to perform any change without drastically afecting the meaning. This opens the avenue
for character-level modifications and, indeed, such solution (SINAI) provided the best results in manual
evaluation.</p>
      <p>We also need to acknowledge the limitations of the evaluation setup. It aims to predict the likelihood
of an AE fooling the victim classifier and transmitting the intended message, as encoded through
BODEGA score. However, human readers are the intended recipients of misinformation, and they are
also able to refuse to engage with a message that seems suspicious, artificial or distorted, e.g. due to use
of letter with non-standard shapes. Thus, the success of AEs will also depend on the visual appearance
of the manipulated content, which is not directly evaluated in the current setup. Quantifying this efect
would be challenging in manual evaluation and even more so in an automatic setup.</p>
      <p>In any case, the results that we do have leave one thing clear: popular architectures for text
classification are very vulnerable to attack with AEs. While the adversarially-trained model posed slightly
harder challenge, ultimately AEs were found for nearly all cases in these scenarios as well.</p>
      <p>How can we protect the real-world deployments of text classifiers against such attack? The first
barrier can be established by limiting the access to the victim model. We can see from the results that
the advance over previous state of the art was accompanied by a raise in the number of queries sent, well
into hundreds and thousands for each generated AE sample. Nevertheless, all machine-learning-based
solutions for content filtering should be only deployed after a thorough analysis of their adversarial
robustness.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In the InCrediblAE shared task, six teams participated with various solutions, both operating with the
word-level and character-level changes. The participants’ approaches (and two reference solutions)
were evaluated using five misinformation-detection scenarios and three victim models.</p>
      <p>In total, 53,544 text modifications were considered and automatically assessed in terms of classifier
confusion, meaning preservation and character similarity. The submitted solutions easily outperformed
previous work in all of the tested scenarios. The manual evaluation highlighted the special role of the
fact-checking tasks and the eficacy of character replacement in performing modification imperceptible
to humans.</p>
      <p>We hope that the combined efort of the participants and organisers of the InCrediblAE shared task
will succeed in both highlighting the importance of robustness testing and showcasing the best solutions.
To facilitate this outcome, the code and resources necessary for performing the automatic evaluation
remain openly available.9</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>The work of P. Przybyła is part of the ERINIA project, which has received funding from the European
Union’s Horizon Europe research and innovation programme under grant agreement No 101060930.
This work has been also partially funded by the European Commission under contract numbers
HE101070278 and ISF-101080090. Views and opinions expressed are however those of the author(s) only
and do not necessarily reflect those of the funders. Neither the European Union nor the granting</p>
      <sec id="sec-7-1">
        <title>9https://github.com/piotrmp/BODEGA</title>
        <p>authority can be held responsible for them. We also acknowledge support from Departament de Recerca
i Universitats de la Generalitat de Catalunya (ajuts SGR-Cat 2021) and from Maria de Maeztu Units of
Excellence Programme CEX2021-001195-M, funded by MCIN/AEI /10.13039/501100011033.
P. G. Allen, RoBERTa: A Robustly Optimized BERT Pretraining Approach (2019). URL: https:
//arxiv.org/abs/1907.11692v1. doi:10.48550/arxiv.1907.11692. arXiv:1907.11692.
[28] L. Li, R. Ma, Q. Guo, X. Xue, X. Qiu, BERT-ATTACK: Adversarial Attack Against BERT Using
BERT, in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language
Processing (EMNLP), Association for Computational Linguistics, 2020, pp. 6193–6202. URL: https:
//aclanthology.org/2020.emnlp-main.500. doi:10.18653/v1/2020.emnlp-main.500.
[29] T. Sellam, D. Das, A. Parikh, BLEURT: Learning Robust Metrics for Text Generation, in:
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association
for Computational Linguistics, Online, 2020, pp. 7881–7892. URL: https://aclanthology.org/2020.
acl-main.704. doi:10.18653/v1/2020.acl-main.704.
[30] V. I. Levenshtein, Binary codes capable of correcting deletions, insertions, and reversals, Soviet</p>
        <p>
          Physics Doklady 10 (1966) 707–710.
[31] Y. Mu, M. Jin, C. Grimshaw, C. Scarton, K. Bontcheva, X. Song, Vaxxhesitancy: A dataset for
studying hesitancy towards covid-19 vaccination on twitter, in: Proceedings of the International
AAAI Conference on Web and Social Media, volume 17, 2023, pp. 1052–1062.
[32] D. Wilby, T. Karmakharm, I. Roberts, X. Song, K. Bontcheva, GATE Teamware 2: An
opensource tool for collaborative document classification annotation, in: D. Croce, L. Soldaini
(Eds.), Proceedings of the 17th Conference of the European Chapter of the Association for
Computational Linguistics: System Demonstrations, Association for Computational
Linguistics, Dubrovnik, Croatia, 2023, pp. 145–151. URL: https://aclanthology.org/2023.eacl-demo.17.
doi:10.18653/v1/2023.eacl-demo.17.
[33] J. Valle Aguilera, A. J. Gutiérrez Megías, S. M. Jiménez Zafra, L. A. Ureña López, E. Martínez Cámara,
SINAI at CheckThat! 2024: Stealthy character-level adversarial attacks using homoglyphs and
search, iterative, in: [
          <xref ref-type="bibr" rid="ref31">45</xref>
          ], 2024.
[34] C. Roadhouse, M. Shardlow, A. Williams, MMU NLP at CheckThat! 2024: Homoglyphs are
adversarial attacks, in: [
          <xref ref-type="bibr" rid="ref31">45</xref>
          ], 2024.
[35] H. He, Y. Song, D. Massey, Palöri at CheckThat! 2024 shared task 6: Glota - combining glove
embeddings with roberta for adversarial attack, in: [
          <xref ref-type="bibr" rid="ref31">45</xref>
          ], 2024.
[36] W. Lewoniewski, P. Stolarski, M. Stróżyna, E. Lewańska, A. Wojewoda, E. Księżniak, M. Sawiński,
OpenFact at CheckThat! 2024: Combining multiple attack methods for efective adversarial text
generation, in: [
          <xref ref-type="bibr" rid="ref31">45</xref>
          ], 2024.
[37] M. Alzantot, Y. Sharma, A. Elgohary, B.-J. Ho, M. Srivastava, K.-W. Chang, Generating Natural
Language Adversarial Examples, in: Proceedings of the 2018 Conference on Empirical Methods in
Natural Language Processing, Association for Computational Linguistics, Brussels, Belgium, 2018,
pp. 2890–2896. URL: https://aclanthology.org/D18-1316. doi:10.18653/v1/D18-1316.
[38] N. Mrkšić, D. Ó. Séaghdha, B. Thomson, M. Gasic, L. M. R. Barahona, P.-H. Su, D. Vandyke, T.-H.
        </p>
        <p>Wen, S. Young, Counter-fitting word vectors to linguistic constraints, in: Proceedings of the 2016
Conference of the North American Chapter of the Association for Computational Linguistics:
Human Language Technologies, 2016, pp. 142–148.
[39] D. Jin, Z. Jin, J. T. Zhou, P. Szolovits, Is BERT Really Robust? A Strong Baseline for Natural
Language Attack on Text Classification and Entailment, in: The Thirty-Fourth AAAI Conference
on Artificial Intelligence, AAAI 2020„ AAAI Press, 2020, pp. 8018–8025. URL: https://ojs.aaai.org/
index.php/AAAI/article/view/6311.
[40] J. Morris, E. Lifland, J. Y. Yoo, J. Grigsby, D. Jin, Y. Qi, TextAttack: A Framework for Adversarial
Attacks, Data Augmentation, and Adversarial Training in NLP, in: Proceedings of the 2020
Conference on Empirical Methods in Natural Language Processing: System Demonstrations,
Association for Computational Linguistics, Online, 2020, pp. 119–126. URL: https://aclanthology.
org/2020.emnlp-demos.16. doi:10.18653/v1/2020.emnlp-demos.16.
[41] D. Li, Y. Zhang, H. Peng, L. Chen, C. Brockett, M.-T. Sun, W. B. Dolan, Contextualized perturbation
for textual adversarial attack, in: Proceedings of the 2021 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021,
pp. 5053–5069.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>A. Annotation Guidelines</title>
      <p>InCrediblAE Shared Task Annotation All participants who submit entries will have their
submissions manually evaluated for semantic similarity. To aid in this human evaluation process, participants
in the shared tasks will be required to manually assess a selection of the submitted examples.
Manual Scoring We will gather assessments regarding the semantic similarities between attack
samples and the original samples. Participants in the shared task are requested to dedicate approximately
60 mins to this manual evaluation (i.e., 100 samples per participant), which will be conducted using this
online annotation tool. Judges will rate the sample pairs based on the following scale: (a). Preserve the
Semantic Meaning, (b). Change the Semantic Meaning, (c). No sense.</p>
      <p>Labels and Definitions (a). Preserve the Semantic Meaning: This label is used when the semantic
content of the attack sample closely aligns with that of the original sample. Participants should use this
label if the meaning, context, and intent of the compared texts remain essentially unchanged, indicating
that the attack sample has efectively maintained the core message of the original. For example:
Text Original: middle-earth. middle-earth is the fictional setting of much of british writer
j. r. r. tolkien ’s legendarium . middle-earth is a setting.</p>
      <p>Text Modified: middle-earth. middle-earth is the fictional background of much of british
writer j. r. r. tolkien ’s legendarium . middle-earth is a setting.</p>
      <p>(b). Change the Semantic Meaning: Participants should apply this label when there is a noticeable
alteration in the semantic content between the original and the attack sample. This label indicates that
while the attack sample may be related or similar in some aspects to the original, it diverges enough in
meaning or intent to be considered distinct or modified. For example, name entities (including Year,
Name, Location, etc.) are changed in the modified text. For example:</p>
      <p>Text Original: carol danvers. the character has also been known as binary , warbird , and
captain marvel at various points in her history , and has been featured in other marvel
licensed products including video games , animated television series , and merchandise
such as trading cards . carol danvers is a real person.</p>
      <p>Text Modified: carol danvers. the character has also been known as binary , warbird , and
captain marvel at various points in her history , and has been featured in other marvel
licensed products including video games , animated television series , and merchandise
such as trading cards . carol danvers is a fictional person.</p>
      <p>(c). No sense: The sentence does not make any sense.</p>
      <p>Confidence Scores Please indicate how confident you are in your assigned class in the Confidence
Row. The confidence scores range from 1 to 5.</p>
      <sec id="sec-8-1">
        <title>A.1. Information Sheet for InCrediblAE Participants</title>
        <p>You are invited to participate in this research by contributing to the evaluation of the semantic similarity
of adversarial examples. It is important for you to understand the goals of the task and what your
participation will involve. Please take the time to read the following information. Please, ask us if there
is anything that is not clear or if you would like more information. Thank you very much for your time.
What is the task’s purpose? The aim of this research is to evaluate the robustness of the text
classifier in adversarial attacks (a detailed description of the task can be found here: https://checkthat.
gitlab.io/clef2024/task6/). In this manual evaluation, we will gather assessments regarding the semantic
similarities between attack samples and the original samples. Participants in the shared task are
requested to dedicate approximately 8 hours to this manual evaluation, which will be conducted using
an online tool. Judges will rate the sample pairs based on the following scale: 3 (Preserve the Semantic
Meaning), 2 (Change the Semantic Meaning), and 1 (No sense).</p>
        <p>Your participation, what it involves and why we are grateful It is up to you to decide whether
or not you want to participate in this annotation task. If you do decide to support us in the project, you
will be given this information sheet to keep (and be asked to sign a separate consent form). You can
still withdraw at any time without any consequences and without giving any reason. The entire data
collected from withdrawn participants will be destroyed immediately, and no personal information
will be kept. If participants submitted a solution to the shared task that also means their submission to
the shared tasks will not be manually scored. If you wish to withdraw, please contact Dr Xingyi Song
(details in section 10). You will be asked to annotate adversarial examples into 4 categories using the
GATE Teamware Platform (https://annotate.gate.ac.uk/).</p>
      </sec>
      <sec id="sec-8-2">
        <title>What are the possible advantages, disadvantages and risks of being involved? Participating</title>
        <p>in this evaluation will support our research on the development of a more accurate assessment of the
robustness of text classifiers and efectiveness of the adversarial attack methods. No major disadvantages
or risks are foreseen, however, it is worth mentioning that the content being annotated may case distress.
You will be in charge of selecting the content to be annotated, therefore, you are free to only select
content that you are comfortable with. If, at any part of the experiment you feel uncomfortable with
the content you are accessing, please talk to one of the responsible researchers.</p>
        <p>Will my involvement be kept confidential? All the information that we collect from you and
about you during the course of the research will be kept strictly confidential and will only be accessible
to members of the research team. You will not be able to be identified in any reports or publications
unless you have given your explicit consent for this on your participant consent form.</p>
      </sec>
      <sec id="sec-8-3">
        <title>What is the legal basis for processing my personal data? According to data protection legislation,</title>
        <p>we are required to inform you that the legal basis we are applying in order to process your personal
data is that ‘processing is necessary for the performance of a task carried out in the public interest’
(Article 6(1)(e)). Further information can be found in the University of Shefield’s Privacy Notice, which
is available online under https://www.shefield.ac.uk/govern/data-protection/privacy/general.</p>
      </sec>
      <sec id="sec-8-4">
        <title>What will happen to the data collected in this study? The data gathered during this scoring</title>
        <p>task will be used to assess the efectiveness of adversarial examples generation. Since this data will
also benefit other researchers, we plan to release a version of the annotated dataset. You will not be
identified and your scoring will be aggregated with multiple other annotators. In the case that you
provided us with your e-mail address at the beginning of the scoring task, we will destroy it after the
scoring task is finished. The University of Shefield will act as the Data Controller for this study.
Who is organising and funding the research? This study is organised jointly by Universitat
Pompeu Fabra and the University of Shefield at the CheckThat! Lab at CLEF 2024. Universitat Pompeu
Fabra is funded by the European Union’s Horizon Europe research and innovation programme under
grant agreement No 101060930. The University of Shefield is funded by the the UK’s innovation agency
(Innovate UK) grant 10039055 (approved under the Horizon Europe Programme as vera.ai EU grant
agreement 101070093).</p>
        <p>Who has ethically reviewed this study? This project has been ethically approved via the University
of Shefield’s Ethics Review Procedure, as administered by the Computer Science Department.
What if something goes wrong and I wish to complain about the research? If you have any
complaints, either from the researcher or something occurring during or following your participation
in the project (e.g. a reportable serious adverse event), please contact Dr. Xingyi Song (contact
details in section 10). Should you feel your complaint has not been handled to your satisfaction, you
can also contact the Head of Department at the University of Shefield, Professor Heidi Christensen
(heidi.christensen@shefield.ac.uk) who will then escalate the complaint through the appropriate
channels. If the complaint relates to how your personal data has been handled, information about
how to raise a complaint can be found in the University’s Privacy Notice: https://www.shefield.ac.uk/
govern/data-protection/privacy/general.</p>
        <p>Contact for further information Details of who you should contact if you wish to obtain further
information are as follows: Dr. Xingyi Song, Department of Computer Science, University of Shefield,
211 Portobello, Shefield S1 4DP, UK. E-mail: x.song@shefield.ac.uk Telephone: +44 114 222 18577.</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>B. COVID-19 Misinformation Dataset Examples</title>
      <p>User Text
COVID-19 the average diameter of the virus particles is
around 120 nm (.12 m). Any mask including N95 masks
can’t filter this particle size, to filter it completely would
prevent the subject being able to breath. Therefore, the
mask are just the new normal, a fashion statement.
A #COVID19 particle is about 1 to 4 microns, an N95
will block 95% of tiny air particles, down to 0.3 microns,
and surgical masks aren’t efective at blocking particles
smaller than 100 microns. Hence surgical masks cannot
stop the spread of #COVID19.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Tucker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Guess</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Barberá</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Vaccari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sanovich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Stukal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Nyhan</surname>
          </string-name>
          , Social Media, Political Polarization, and
          <article-title>Political Disinformation: A Review of the Scientific Literature</article-title>
          ,
          <source>Technical Report, Hewlett Foundation</source>
          ,
          <year>2018</year>
          . URL: https://hewlett.org/library/ social
          <article-title>-media-political-polarization-political-disinformation-review-scientific-literature/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>S. van der Linden</surname>
          </string-name>
          ,
          <article-title>Misinformation: susceptibility, spread, and interventions to immunize the public</article-title>
          ,
          <source>Nature Medicine 2022 28:3</source>
          <volume>28</volume>
          (
          <year>2022</year>
          )
          <fpage>460</fpage>
          -
          <lpage>467</lpage>
          . URL: https://www.nature.com/articles/ s41591-022-01713-6. doi:
          <volume>10</volume>
          .1038/s41591-022-01713-6.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kiesel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mestre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shukla</surname>
          </string-name>
          , E. Vincent,
          <string-name>
            <given-names>P.</given-names>
            <surname>Adineh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Corney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          , M. Potthast, SemEval
          <article-title>-2019 Task 4: Hyperpartisan News Detection</article-title>
          ,
          <source>in: Proceedings of the 13th International Workshop on Semantic Evaluation</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Minneapolis, Minnesota, USA,
          <year>2019</year>
          , pp.
          <fpage>829</fpage>
          -
          <lpage>839</lpage>
          . URL: https://aclanthology.org/S19-2145. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>S19</fpage>
          -2145.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G. da San</given-names>
            <surname>Martino</surname>
          </string-name>
          , A.
          <string-name>
            <surname>Barrón-Cedeño</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wachsmuth</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Petrov</surname>
          </string-name>
          , P. Nakov, SemEval-2020
          <source>Task</source>
          <volume>11</volume>
          :
          <article-title>Detection of Propaganda Techniques in News Articles</article-title>
          ,
          <source>in: Proceedings of the Fourteenth Workshop on Semantic Evaluation (SemEval-2020)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1377</fpage>
          -
          <lpage>1414</lpage>
          . URL: http://propaganda. qcri.org/annotations/definitions.htmlhttp://arxiv.org/abs/
          <year>2009</year>
          .02696. arXiv:
          <year>2009</year>
          .02696.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <source>Overview of the 7th Author Profiling Task at PAN</source>
          <year>2019</year>
          :
          <article-title>Bots and Gender Profiling</article-title>
          , in: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>D. E.</given-names>
          </string-name>
          <string-name>
            <surname>Losada</surname>
          </string-name>
          , H. Müller (Eds.),
          <article-title>CLEF 2019 Labs and Workshops, Notebook Papers</article-title>
          .
          <source>CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Cocarascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <source>The FEVER2</source>
          .
          <article-title>0 Shared Task</article-title>
          ,
          <source>in: Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Singhal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Paudel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Thota</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kumarswamy</surname>
          </string-name>
          , G. Stringhini, S. Nilizadeh, SoK: Content Moderation in Social Media, from Guidelines to Enforcement, and Research to Practice,
          <source>in: The 8th IEEE European Symposium on Security and Privacy (EuroS&amp;P</source>
          <year>2023</year>
          ), IEEE,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2206.14855v2. doi:
          <volume>10</volume>
          .48550/arxiv.2206.14855. arXiv:
          <volume>2206</volume>
          .
          <fpage>14855</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaremba</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bruna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <article-title>Intriguing properties of neural networks</article-title>
          ,
          <source>arXiv: 1312.6199</source>
          (
          <year>2013</year>
          ). URL: https://arxiv.org/abs/1312.6199v4. doi:
          <volume>10</volume>
          .48550/arxiv.1312.6199. arXiv:
          <volume>1312</volume>
          .
          <fpage>6199</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W. E.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. Z.</given-names>
            <surname>Sheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Alhazmi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Adversarial Attacks on Deep-learning Models in Natural Language Processing</article-title>
          ,
          <source>ACM Transactions on Intelligent Systems and Technology (TIST) 11</source>
          (
          <year>2020</year>
          ). URL: https://dl.acm.org/doi/10.1145/3374217. doi:
          <volume>10</volume>
          .1145/3374217.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <article-title>The clef-2024 checkthat! lab: Check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>449</fpage>
          -
          <lpage>458</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Piskorski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2024 CheckThat! Lab: Check-worthiness, subjectivity, persuasion, roles, authorities and adversarial robustness</article-title>
          , in: L.
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mulhem</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Quénot</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Soulier</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Di Nunzio</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Galuščáková</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ),
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shvets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <source>Verifying the Robustness of Automatic Credibility Assessment, arXiv preprint arXiv:2303.08032</source>
          (
          <year>2023</year>
          ). URL: https://arxiv.org/abs/2303.08032v1. arXiv:
          <volume>2303</volume>
          .
          <fpage>08032</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Zeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          , M. Sun,
          <article-title>OpenAttack: An Open-source Textual Adversarial Attack Toolkit, in: ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th</article-title>
          <source>International Joint Conference on Natural Language Processing, Proceedings of the System Demonstrations, Association for Computational Linguistics (ACL)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>363</fpage>
          -
          <lpage>371</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .acl-demo.
          <volume>43</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2021</year>
          .ACL-DEMO.
          <year>43</year>
          . arXiv:
          <year>2009</year>
          .09191.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B. D.</given-names>
            <surname>Horne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adali</surname>
          </string-name>
          , This Just In:
          <article-title>Fake News Packs a Lot in Title, Uses Simpler, Repetitive Content in Text Body, More Similar to Satire than Real News</article-title>
          ,
          <source>in: Proceedings of the 2nd International Workshop on News and Public Opinion at ICWSM, Association for the Advancement of Artificial Intelligence</source>
          ,
          <year>2017</year>
          . URL: http://arxiv.org/abs/1703.09398. arXiv:
          <volume>1703</volume>
          .
          <fpage>09398</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <article-title>Capturing the Style of Fake News</article-title>
          ,
          <source>in: Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20)</source>
          , volume
          <volume>34</volume>
          , AAAI Press, New York, USA,
          <year>2020</year>
          , pp.
          <fpage>490</fpage>
          -
          <lpage>497</lpage>
          . URL: https://aaai.org/ojs/index.php/AAAI/article/view/5386. doi:
          <volume>10</volume>
          .1609/aaai. v34i01.
          <fpage>5386</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T. J.</given-names>
            <surname>Smith</surname>
          </string-name>
          , Propaganda:
          <string-name>
            <given-names>A Pluralistic</given-names>
            <surname>Perspective</surname>
          </string-name>
          , Praeger,
          <year>1989</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>L.</given-names>
            <surname>Graves</surname>
          </string-name>
          ,
          <source>Understanding the Promise and Limits of Automated Fact-Checking</source>
          ,
          <source>Technical Report</source>
          , Reuters Institute, University of Oxford,
          <year>2018</year>
          . URL: https://reutersinstitute.politics.ox.ac.uk/sites/ default/files/2018-02/graves{_}factsheet{_}180226FINAL.pdf. arXiv:arXiv:
          <fpage>1011</fpage>
          .
          <year>1669v3</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Cocarascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <article-title>The Fact Extraction and VERification (FEVER) Shared Task</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Fact Extraction and VERification (FEVER)</source>
          ,
          <year>2018</year>
          . arXiv:
          <year>1811</year>
          .10971v1.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>M.</given-names>
            <surname>Al-Sarem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Boulila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Al-Harby</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Qadir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Alsaeedi</surname>
          </string-name>
          ,
          <article-title>Deep learning-based rumor detection on microblogging platforms: A systematic review</article-title>
          ,
          <source>IEEE Access 7</source>
          (
          <year>2019</year>
          )
          <fpage>152788</fpage>
          -
          <lpage>152812</lpage>
          . doi:
          <volume>10</volume>
          . 1109/ACCESS.
          <year>2019</year>
          .
          <volume>2947855</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>S.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J</given-names>
            .
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ciravegna</surname>
          </string-name>
          ,
          <article-title>Neural language model based training data augmentation for weakly supervised early rumor detection</article-title>
          ,
          <source>in: Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, ASONAM</source>
          <year>2019</year>
          ,
          <article-title>Association for Computing Machinery</article-title>
          , Inc,
          <year>2019</year>
          , pp.
          <fpage>105</fpage>
          -
          <lpage>112</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3341161.3342892. doi:
          <volume>10</volume>
          . 1145/3341161.3342892. arXiv:
          <year>1907</year>
          .07033.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scarton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          ,
          <article-title>Categorising fine-to-coarse grained misinformation: An empirical study of the covid-19 infodemic</article-title>
          , in
          <source>: Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>556</fpage>
          -
          <lpage>567</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Heppell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scarton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <article-title>A large-scale comparative study of accurate covid-19 information versus misinformation</article-title>
          ,
          <source>arXiv preprint arXiv:2304.04811</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hochreiter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Long</given-names>
            <surname>Short-Term</surname>
          </string-name>
          <string-name>
            <surname>Memory</surname>
          </string-name>
          ,
          <source>Neural Computation</source>
          <volume>9</volume>
          (
          <year>1997</year>
          )
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          . doi:
          <volume>10</volume>
          .1162/neco.
          <year>1997</year>
          .
          <volume>9</volume>
          .8.1735.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          , Association for Computational Linguistics,
          <year>2018</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: http://arxiv.org/abs/
          <year>1810</year>
          .04805. arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Louf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , P. von Platen, C. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gugger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Drame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lhoest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Rush</surname>
          </string-name>
          , Transformers:
          <article-title>State-of-the-Art Natural Language Processing</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>38</fpage>
          -
          <lpage>45</lpage>
          . URL: https://www.aclweb.org/anthology/2020.emnlp-demos.
          <volume>6</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>I.</given-names>
            <surname>Loshchilov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hutter</surname>
          </string-name>
          , Decoupled Weight Decay Regularization,
          <source>in: 7th International Conference on Learning Representations, ICLR</source>
          <year>2019</year>
          ,
          <article-title>New Orleans</article-title>
          , LA, USA,
          <year>2019</year>
          . URL: https://openreview. net/forum?id=Bkg6RiCqY7. arXiv:
          <volume>1711</volume>
          .
          <year>05101v3</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>B.</given-names>
            <surname>Demirok</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mergen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Oz</surname>
          </string-name>
          , TurQUaz at CheckThat! 2024:
          <article-title>Creating adversarial examples using genetic algorithm</article-title>
          ,
          <source>in: [45]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>D.</given-names>
            <surname>Guzman Piedrahita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fazla</surname>
          </string-name>
          , L. Krauter, TextTrojaners at CheckThat! 2024:
          <article-title>Robustness of credibility assessment with adversarial examples through beamattack</article-title>
          ,
          <source>in: [45]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>" why should i trust you?" explaining the predictions of any classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>G.</given-names>
            <surname>Faggioli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Galuščáková</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . García Seco de Herrera (Eds.), Working Notes of CLEF 2024 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2024</year>
          , Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>