<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Zorros at CheckThat! 2022: Ensemble Model for Identifying Relevant Claims in Tweets</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nicu Buliga</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Madalina Raschip</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bitdefender</institution>
          ,
          <addr-line>Iasi</addr-line>
          ,
          <country country="RO">Romania</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Computer Science, "Alexandru Ioan Cuza" University of Iasi</institution>
          ,
          <addr-line>Iasi</addr-line>
          ,
          <country country="RO">Romania</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes the system used by Zorros team in the CLEF2022 CheckThat! Lab for Task 1 on identifying relevant claims in tweets. Task 1 was divided into four subtasks, which try to detect if the tweets are worth fact-checking (1A), contain verifiable factual claims (1B), are harmful to society (1C) and are attention-worthy (1D). For each subtask, we proposed diferent models based on pre-trained transformer models that helped us achieve the first position for subtasks 1C and 1D, the second position for subtask 1A, and the fifth position for subtask 1B.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;check-worthiness</kwd>
        <kwd>COVID-19</kwd>
        <kwd>transformer models</kwd>
        <kwd>ensemble</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and T5. Those pre-trained models can be used to transfer knowledge to other NLP tasks,
leading to significant improvements. Diferent pre-trained BERT [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and RoBERTa [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] models
are used in ensemble models to solve the subtasks 1A-1D.
      </p>
      <p>The rest of the paper is organized as follows: section 2 - Related Work, section 3 - Problem
Description, section 4 - Methodology, where the models used are described, section 5 - Evaluation
and Results from the competition and section 6 - Conclusion and Future Work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Fake news detection has gained a lot of attention in the last years, especially during the
worldwide COVID-19 pandemic, because of the misinformation spreading about COVID-19 on
social media and there is a need of platforms to prevent it. Thus, many systems have been tested
and used for detecting fake news, but they can not classify information accurately, because of
their inability to fully understand the data.</p>
      <p>
        Some of the most recent deep learning models used for fake news detection tasks are
described below. The paper [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] presents a hybrid Neural Network architecture that combines the
capabilities of CNN and LSTM. Two diferent dimensionality reduction approaches, Principle
Component Analysis and Chi-Square are used in order to reduce the dimensionality of the
feature vectors before passing them to the classifier. To develop the idea, the authors acquired a
dataset from the Fake News Challenges website which has four types of stances: agree, disagree,
discuss, and unrelated. Their results show a 20% improvement in the F1-Score. Another model
is described in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The authors used a transformer model for fake news classification of a
specific domain dataset, and included human justification and metadata for added performance.
They have used multiple BERT models with shared weights between them to handle various
inputs.
      </p>
      <p>
        Related research areas are check-worthiness and credibility assessment of tweets. There
are many approaches in literature for identifying check-worthiness in social media, starting
from working with features like TF-IDF representations [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] until more recently used word
embeddings from transformers. For example, the paper [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] analyzes diferent classifiers and
uses diferent features like TF-IDF representations, part of speech tags, sentiment scores, and
entity types to detect check-worthy factual claims in presidential debates.
      </p>
      <p>CLEF has been organizing CheckThat! Labs since 2018. The best model [10] from 2018
for check-worthiness in political claims used a multilayer perceptron and many features like
averaged word embeddings and bag-of-words representations. In the last two editions, with the
emergence of the COVID-19 pandemic, the competition considered the task of check-worthiness
of tweets about COVID-19. Also, transformer-based models have begun to be used often. In the
last year, for the check-worthiness of tweets task the best approach [11] used several pre-trained
transformer models, BERTweet [12] giving the best results on the development set.</p>
      <p>Some recent works for credibility assessment are presented below. In [13], a semi-supervised
ranking model is described to automatically evaluate the credibility of a tweet. In [14], a large
multilingual dataset for fact checking is introduced along with several automated fact checking
models based on transformers.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Problem Description</title>
      <p>The Task 1 of the competition was separated into four subtasks, all of them related to the
COVID-19 topic. The first three tasks are binary classification tasks, and the last one is a
multiclass classification task. Every subtask had its own dataset, splitted into train, dev and test
sets as shown in Table 1.</p>
      <p>The label distributions are shown in Figure 1 for subtasks 1A, 1B, 1C and in Figure 2 for
subtask 1D. We can clearly observe that for subtasks 1A, 1C and 1D the data are very unbalanced,
so we have to use diferent techniques for balancing the data.
• Subtask 1A: Check-worthiness of tweets</p>
      <p>Given a tweet, the task is to predict whether it is worth fact-checking, so it has two labels:</p>
      <p>Yes and No.
• Subtask 1B: Verifiable factual claims detection</p>
      <p>For this subtask, we have to predict whether a tweet contains a verifiable factual claim.
• Subtask 1C: Harmful tweet detection</p>
      <p>We have to predict whether a tweet is harmful to society.
• Subtask 1D: Attention-worthy tweet detection</p>
      <p>This subtask has nine class labels and we have to predict whether a tweet should get the
attention of policy makers and why. The labels are:
– No, not interesting
– Yes, asks question
– Yes, blame authorities
– Yes, calls for action
– Yes, harmful
– Yes, contains advice
– Yes, discusses action taken
– Yes, discusses cure
– Yes, other</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>The proposed approach for solving the subtasks 1A-1D consists of four main steps: text
preprocessing, tokenization for transformer based models, selection of the model architecture and the
construction of the ensemble model. Each of them is described below.</p>
      <sec id="sec-4-1">
        <title>4.1. Text Preprocessing</title>
        <p>Given the small dataset for each task and the irrelevance of some tokens in the tweets for
training our model, we performed the following modifications on raw tweets:</p>
        <sec id="sec-4-1-1">
          <title>1. Lowercased the text of the tweet</title>
          <p>2. Replaced all shorts, like don’t to their normal form, do not
3. Removed all URLs, TAGs and non alphanumeric characters
4. Removed all stand-alone numbers</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Tokenization</title>
        <p>To pass the tweets to the pre-trained models like BERT and RoBERTa, we need to convert every
tweet into a list of tokens based on the vocabulary of the model and then find the tokens ids
and the attention masks. For this step, we used the respective tokenizer of each model because
it already knows the accepted structure of the model and can easily convert the tweets.</p>
        <p>The sentences are grouped into batches, so the tweets in the same batch must have the
same number of tokens. To satisfy this condition, we used the parameters of the tokenizer by
specifying a maximum length based on tokens list length distribution (100 was the best length
in experiments). Therefore, shorter tweets will be padded with the same predefined token and
longer tweets will be truncated.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Model Architecture</title>
        <p>We have used only the encoder block of pre-trained BERT and RoBERTa models as the core of
the final model, which can ofer really good performance on NLP tasks. On top of that, we need
a classification header for predicting, because until now we have only tweets embeddings. We
need a binary classification model for subtasks 1A, 1B, 1C and a slightly diferent model for
subtask 1D, because it has more than two labels. So, for the first three subtasks, we added an
output neuron with the sigmoid as an activation function on top of the pre-trained model with
a threshold at 0.5, giving us the probability of a tweet being in class YES. The loss function is
Binary Cross Entropy and the optimizer is Adam with a weight decay of 0.01 and a learning
rate of 2 · 10− 5. The classification head has a dropout layer with a rate of 0.5. Dropout was
applied to avoid over-fitting. The model architecture is given in Figure 3.</p>
        <p>For the final subtask 1D, we have as a classification header nine neurons with softmax as
an activation function and Cross Entropy as a loss function. The predicted output gives us a
probability distribution over all classes. The optimizer respects the same parameters from the
ifrst model, and the dropout layer also has a rate of 0.5. The model architecture for subtask 1D
is given in Figure 4.</p>
        <p>We fine-tuned these models for every subtask on the train set consisting of a concatenation
between the train set and the test set. For testing models performance we have used the dev
set. For training, we chose 20 epochs with data grouped into batches of 16 tweets. To deal
with the unbalanced datasets, we have used weights for classes in the loss functions, essentially
assigning a higher weight to the loss of the minor classes.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Fine-tuned BERT and RoBERTa Models</title>
        <sec id="sec-4-4-1">
          <title>We fine-tuned the following ten models for every subtask:</title>
          <p>4.4.1. TweetEval based models
TweetEval [15] is a pre-trained RoBERTa-base model, further trained on ∼ 58 tweets,
randomly collected. The result of this step is a Twitter-domain adapted version of RoBERTa.</p>
          <p>We also used a selection of five TweetEval models, each fine tuned for a specific task: Irony
Detection[16], Ofensive Language Identification[ 17], Emotion Recognition[18], Hate Speech
Detection[19], Sentiment Analysis[20].
4.4.2. BERTweet models
BERTweet [12] is the first public large-scale language model pre-trained for English Tweets.
BERTweet is trained based on the RoBERTa pre-training procedure. The corpus used to pre-train
BERTweet consists of 850 English Tweets (16 word tokens ∼ 80), containing 845
Tweets streamed from 01/2012 to 08/2019 and 5 Tweets related to the COVID-19 pandemic.</p>
          <p>We used three versions of this model:
• BERTweet Large, the large version of the model, pre-trained on 873 English tweets
(cased)
• BERTweet COVID-19 Base Cased, the base version of the model, additionally trained on
23 COVID-19 English tweets (cased)
• BERTweet COVID-19 Base Uncased, the base version of the model, additionally trained
on 23 COVID-19 English tweets (uncased)
4.4.3. COVID-Twitter-BERT v2
It is a BERT-large-uncased model, pre-trained on a corpus of messages from Twitter about
COVID-19. This model is identical to COVID-Twitter-BERT v1 [21] but was trained on more
data, resulting in higher downstream performance.</p>
          <p>The first version of the model was trained on 160 tweets collected between January 12 and
April 16, 2020, containing at least one of the keywords "wuhan", "ncov", "coronavirus", "covid",
or "sars-cov-2". These tweets were filtered and preprocessed to reach a final sample of 22.5
tweets, containing 40.7 sentences and 633 tokens, which were used for training.</p>
        </sec>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Ensemble Model</title>
        <p>For subtasks 1A, 1B and 1C we used an ensemble model to predict the labels. The structure of
the ensemble model is composed of ten BERT and RoBERTa pre-trained and fine-tuned models
described above, with another classification header on top of them, with one neuron and sigmoid
as activation function, predicting the probability of a tweet being in class YES. Essentially, we
took the predictions from every fine-tuned model without the sigmoidal activation function
and feed them to the classification header. Therefore every tweet will have a new feature vector
of length 10, where an element is a raw prediction from one model.</p>
        <p>The model structure can be seen in Figure 5.</p>
        <p>This new model was trained for 50 epochs with a batch size of 8 and a learning rate of 2 · 10− 4,
checking its performance at every epoch and saving the model with the best performance.</p>
        <p>For task 1D, we did not use an ensemble model but the model with the best performance out
of all ten, which was COVID-Twitter-BERT v2 [21].</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation and Results</title>
      <p>Task 1 is a classification task. Classification algorithms can be evaluated using several metrics
including accuracy, precision, recall, and F1-score. For the subtasks 1A and 1C, the organizers
used the F1 measure with respect to the positive class (minor class), for subtask 1B - accuracy
and for 1D - weighted F1 score. We describe next the results obtained for every subtask on the
last test set ofered by the organizers and used for the contest. Tables 2-5 contain the results for
all metrics; the column in bold from each table is the metric used by the organizers to establish
the winners.</p>
      <p>The results for subtask 1A can be found in Table 2. Here we have tested ten transformer-based
models and an ensemble model built from these. We can observe that the ensemble model has
the best performance for every metric except accuracy, which is not that relevant when the
data is very unbalanced. Therefore, this is the model we used for the contest.</p>
      <p>For subtask 1B, we tested only five TweetEval derived models and an ensemble model made
from these, because of the increased computation time. The results are given in Table 3. The
ensemble model yield the best results for this subtask, except for the macro recall, which is only
0.01 smaller than the maximum value of this metric. Therefore, we used the ensemble model
for this subtask in the competition.</p>
      <p>The results for subtask 1C are illustrated in Table 4. The tested models are the same as those
from subtask 1B. Here the ensemble has the best results for two out of five metrics, including
the F1 score for positive class, which was used as evaluation for the contest.</p>
      <p>For subtask 1D, we did not use an ensemble. We tested four models due to increased
computation time. The results are given in Table 5. Here the decision was simple, because COVID
Twitter BERT v2 had the best results on three out of four metrics, including the one used for
evaluation.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and Future Work</title>
      <p>In this paper, we present the results obtained in Task 1 of the CLEF 2022 CheckThat! lab.
Diferent models based on pre-trained BERT encoders and fine-tuned BERT models ofered
really good results. For the first three subtasks we used ensemble models, while for subtask 1D
a fine-tuned BERT model was enough. We obtained first place for subtasks 1C and 1D, second
place for subtask 1A and fifth place for subtask 1B. The experimental results show the power
of the existing pre-trained models, no longer being necessary to retrain a model from scratch,
saving a lot of time on computation.</p>
      <p>For future work, we will focus on learning more features from a sentence, making diferent
combinations from BERT layers, like pooling the last four layers, or concatenating them to
obtain a higher understanding of the semantic in a sentence.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgement</title>
      <p>This paper is partially supported by the Competitiveness Operational Programme Romania
under project number SMIS 124759 - RaaS-IS (Research as a Service Iasi).
debates, in: Proceedings of the 24th ACM International on Conference on Information
and Knowledge Management, 2015, pp. 1835–1838.
[10] C. Zuo, A. Karakas, R. Banerjee, A hybrid recognition system for check-worthy claims
using heuristics and supervised learning, in: CEUR workshop proceedings (Vol. 2125),
2018.
[11] J. R. Martinez-Rico, J. Martinez-Romo, L. Araujo, Nlpir@uned at checkthat! 2021:
checkworthiness estimation and fake news detection using transformer models, in: CEUR
workshop proceedings (Vol. 2936), 2021.
[12] D. Q. Nguyen, T. Vu, A. T. Nguyen, BERTweet: A pre-trained language model for English
Tweets, in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language
Processing: System Demonstrations, 2020, pp. 9–14.
[13] A. Gupta, P. Kumaraguru, C. Castillo, P. Meier, Tweetcred: Real-time credibility assessment
of content on twitter, in: International conference on social informatics, 2014, pp. 228–243.
[14] A. Gupta, V. Srikumar, X-fact: A new benchmark dataset for multilingual fact checking,
in: arXiv preprint arXiv:2106.09248, 2021.
[15] F. Barbieri, J. Camacho-Collados, L. Espinosa-Anke, L. Neves, TweetEval:Unified
Benchmark and Comparative Evaluation for Tweet Classification, in: Proceedings of Findings of
EMNLP, 2020.
[16] C. Van Hee, E. Lefever, V. Hoste, Semeval-2018 task 3: Irony detection in english tweets,
in: Proceedings of The 12th International Workshop on Semantic Evaluation, 2018, pp.
39–50.
[17] M. Zampieri, S. Malmasi, P. Nakov, S. Rosenthal, N. Farra, R. Kumar, Semeval-2019
task 6: Identifying and categorizing ofensive language in social media (ofenseval), in:
Proceedings of the 13th International Workshop on Semantic Evaluation, 2019, pp. 75–86.
[18] S. Mohammad, F. Bravo-Marquez, M. Salameh, S. Kiritchenko, Semeval-2018 task 1: Afect
in tweets, in: Proceedings of the 12th international workshop on semantic evaluation,
2018, pp. 1–17.
[19] V. Basile, C. Bosco, E. Fersini, D. Nozza, V. Patti, F. M. Rangel Pardo, P. Rosso, M. Sanguinetti,
SemEval-2019 task 5: Multilingual detection of hate speech against immigrants and women
in Twitter, in: Proceedings of the 13th International Workshop on Semantic Evaluation,
Association for Computational Linguistics, Minneapolis, Minnesota, USA, 2019, pp. 54–63.</p>
      <p>URL: https://www.aclweb.org/anthology/S19-2007. doi:10.18653/v1/S19-2007.
[20] S. Rosenthal, N. Farra, P. Nakov, Semeval-2017 task 4: Sentiment analysis in twitter, in:
Proceedings of the 11th international workshop on semantic evaluation (SemEval-2017),
2017, pp. 502–518.
[21] M. Müller, M. Salathé, P. E. Kummervold, Covid-twitter-bert: A natural language processing
model to analyse covid-19 content on twitter, arXiv preprint arXiv:2005.07503 (2020).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pierri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. L.</given-names>
            <surname>Perry</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. R. DeVerna</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-C. Yang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Flammini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Menczer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Bryden</surname>
          </string-name>
          ,
          <article-title>Online misinformation is linked to early covid-19 vaccination hesitancy and refusal</article-title>
          ,
          <source>Scientific reports 12</source>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Beltrán,</surname>
          </string-name>
          <article-title>The clef-2022 checkthat! lab on fighting the covid-19 infodemic and fake news detection</article-title>
          , in: M.
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Verberne</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Seifert</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Nørvåg</surname>
          </string-name>
          , V. Setty (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer International Publishing, Cham,
          <year>2022</year>
          , pp.
          <fpage>416</fpage>
          -
          <lpage>428</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beltrán</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2022 CheckThat! lab task 1 on identifying relevant claims in tweets</article-title>
          , in: Working Notes of CLEF 2022-
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , CLEF '
          <year>2022</year>
          , Bologna, Italy,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beltrán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Köhler</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2022 CheckThat! lab on fighting the COVID-19 infodemic and fake news detection</article-title>
          ,
          <source>in: Proceedings of the 13th International Conference of the CLEF Association: Information Access Evaluation meets Multilinguality</source>
          , Multimodality, and Visualization, CLEF '
          <year>2022</year>
          , Bologna, Italy,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Umer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Imtiaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ullah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mehmood</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. S.</given-names>
            <surname>Choi</surname>
          </string-name>
          , B.-W. On,
          <article-title>Fake news stance detection using deep learning architecture (cnn-lstm)</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>156695</fpage>
          -
          <lpage>156706</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2020</year>
          .
          <volume>3019735</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dwivedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Patra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Anand</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>A transformer-based architecture for fake news classification</article-title>
          ,
          <source>Social Network Analysis and Mining</source>
          <volume>11</volume>
          (
          <year>2021</year>
          )
          <article-title>39</article-title>
          . URL: https://doi.org/10.1007/s13278-021-00738-y. doi:
          <volume>10</volume>
          .1007/s13278-021-00738-y.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>N.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tremayne</surname>
          </string-name>
          ,
          <article-title>Detecting check-worthy factual claims in presidential</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>