<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>F. Balouchzahi); hlsrekha@gmail.com (H. L. Shashirekha); sidorov@cic.ipn.mx
(G. Sidorov)
~ https://mangaloreuniversity.ac.in/dr-h-l-shashirekha (H. L. Shashirekha); http://www.cic.ipn.mx/~sidorov/
(G. Sidorov)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>MUCIC at CheckThat! 2021: FaDo-Fake News Detection and Domain Identification using Transformers Ensembling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fazlourrahman Balouchzahi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hosahalli Lakshmaiah Shashirekha</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Grigori Sidorov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Computing Research, Instituto Politécnico Nacional</institution>
          ,
          <addr-line>CDMX</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Since the beginning of Covid-19 era in November 2019, the patient growth curve is closely accompanied by the growth of fake news. Therefore, developing tools and models for the detection of fake news from real ones in various domains have become more significant than the earlier days. To address the detection of fake news, in this paper, we, team MUCIC, describe the models submitted to 'Fake News Detection', a shared task organized by CLEF-2021-CheckThat! Lab. This shared task contains two subtasks namely; Fake News Detection of News Articles (Subtask 3A) and Topical Domain Classification of News Articles (Subtask 3B) and both are multi-class text classification tasks. The proposed models have been developed by fine-tuning the three transformer-based language models namely; Roberta, Distilbert, and BERT from HuggingFace using training data and then ensembling them as estimators with majority voting. The proposed models performances evaluated through the evaluation script provided by organizers obtained F1-scores of 0.5309 and 0.8550 for Subtask 3A and Subtask 3B respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Fake News Detection</kwd>
        <kwd>Domain Identification</kwd>
        <kwd>Transformers</kwd>
        <kwd>BERT</kwd>
        <kwd>Roberta</kwd>
        <kwd>Distilbert</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Anonymity is a significant attribute of the cyber world [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] which also provides ample
opportunities to mislead and manipulate peoples’ thoughts and damnify social trust [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Ease of access,
low cost, and swift broadcasting of the information in social media and network is exerting
negative influence on sharing fake news in various domains [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Fake news detection in diferent
domains has received much attention especially after the outbreak of Covid-19 and its efects
on the entire world.
      </p>
      <p>The nature of texts in social media is highly unstructured and noisy, and texts may belong
to various domains as people comment or share messages about various topics. Therefore,
identifying the domain of texts in which news are disseminating in social media is very important
and might help in designing the models/tools to detect fake news. Fake news detection and
domain identification tasks could be modeled as binary Text Classification (TC) problems if
there are only two classes, otherwise they will be modeled as multi-class TC problem.</p>
      <p>
        Hitherto many models based on Machine Learning (ML) and Deep Learning (DL) are
experimented by researchers for TC in general. Of late Transfer Learning (TL) and transformers-based
models are also getting attention due to their eficient performances in various Natural Language
Processing (NLP) and TC tasks [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
      </p>
      <p>Many DL and Neural Network (NN) based models such as CNN, LSTM, BiLSTM, etc. are
considered as best models for many NLP and TC tasks. But, the introduction of transformers
have changed the game and since 2017 transformer1-based models are beating DL models
in many NLP related applications2. Transformers are novel architectures with self-attention
mechanism that are primarily used for NLP. Transformer-based models are able to handle
long-range dependencies and usually are used for sequence-to-sequence NLP tasks3.</p>
      <p>
        NLP researchers are challenged by Conference and Labs of the Evaluation Forum (CLEF)
20214 CheckThat! Lab on Detecting Check-Worthy Claims, Previously Fact-Checked Claims,
and Fake News5 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Fake News Detection (Task 3)6 is one of the three tasks evaluated by the
CheckThat! Lab and this task has two subtasks. Figure 1 illustrates the graphical representation
of subtasks and details of the subtasks are as follows:
• Subtask 3A - Multi-Class Fake News Detection of News Articles (English): is a
multi-class TC task that accepts a given English text and classifies it into one of four
pre-defined categories namely, ‘False’, ‘Partially False’, ‘True’, and ‘Other’. Table 1 gives
the description of labels in Subtask 3A;
• Subtask 3B - Topical Domain Classification of News Articles: is also a multi-class
TC task that further classifies a given fake news into one of six categories representing
six diferent domains namely, ‘Health’, ‘Climate’, ‘Economy’, ‘Crime’, ‘Elections’ and
‘Education’. It is worth to note that classification results of Subtask 3A are incommunicable
to Subtask 3B and only known fake news provided by the organizers are used for Subtask
3B.
      </p>
      <p>
        Many tools and algorithms have been introduced by researchers for TC in general. However,
an algorithm which performs well for a particular dataset may not give the same performance
for other datasets [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Therefore, it is not logical to claim that an algorithm or a model leads
to the same performance for all the datasets. As a result, inspired by H. L. Shashirekha et al.
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and Gundapu S et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], we, team MUCIC, utilized available at HuggingFace7 to develop
transformer-based models, namely, BERT, Distilbert, and Roberta, extended each of these
models with a step of fine-tuning the respective Language Model (LM) and finally ensembled
the extended models with majority voting for fake news detection and domain identification.
1https://pypi.org/project/transformers/
2https://yale-lily.github.io/public/matt_f2018.pdf
3https://towardsdatascience.com/transformers-89034557de14
4http://clef2021.clef-initiative.eu/
5https://sites.google.com/view/clef2021-checkthat/home
6https://sites.google.com/view/clef2021-checkthat/tasks/task-3-fake-news-detection
7https://huggingface.co/
      </p>
      <p>The rest of paper is organized as follows: Section 2 highlights the recent literature and related
works followed by the proposed Ensemble of transformer-based model for fake news detection
and domain identification in Section 3. While Section 4 describes the experiments and results,
Section 5 concludes the paper with future scope.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Alongside the outbreak of Covid-19, fake news detection task has gained more and more
attention due to critical situation the entire globe is facing. Many researchers have developed
many models to combat fake news and avoid its spread thereby. Few of the latest and related
works are briefly discussed in this section. Gundapu et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] have proposed several models
based on ML, DL and TL approaches for the task of fake news detection in English language in
which a given news item is categorized as either ‘fake’ or ‘real’. They conducted experiments
on the dataset provided by ConstraintAI’218 shared task organizers and their ensemble of
transformer-based models outperformed the rest of the models with an F1-score of 0.9855. The
8https://constraint-shared-task-2021.github.io/
dataset as described in detail by Patwa et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] consists of 10,700 texts from various social media
platforms such as Instagram, Twitter, etc. Their proposed transformer-based ensemble model
architecture consists of preprocessing and ensemble model construction. Data is preprocessed by
converting emoji and hashtags to text, removing punctuation, digits, non-ASCII characters, stop
words and stemming. Three transformer-based models namely, BERT, ALBERT, and XLNet are
ensembled in such way that the average of all softmax values from estimators gives probability
of the final predicted label.
      </p>
      <p>
        A multilingual cross-domain dataset containing 5,182 fact-checked news articles related to
Covid-19 is developed by Shahi et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] by collecting articles from 92 fact checking websites
shared during the first five months of 2020. The collected articles include texts in 40 languages,
categorized into 11 classes and then filtered into two categories namely, ‘False’ and ‘Other’. 4,132
texts which belong to ‘False’ category contain false information and the remaining belongs to
‘Other’ category. Authors used Bert-based model as a baseline and obtained an average F1-score
of 0.76 on identifying fake news for the developed dataset.
      </p>
      <p>
        Another work towards fake news detection in Covid-19 domain carried out by Paka et. al [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
includes developing a COVID-19 (Twitter) Fake news (CTF) dataset. CTF is a large-scale text
dataset in Covid-19 domain collected from Twitter containing 21.85M unlabelled tweets along
with 45.26K labeled tweets; out of which 18.55K are labeled as genuine and remaining 26.71K as
fake. Data collection and annotation procedure has been done in four stages namely, Segregating
COVID-19 related tweets, Collecting COVID-19 supporting statements, Filtering genuine and
fake tweets, and Human annotation. The authenticity of the CTF dataset is guaranteed by
fact checking websites such as PolitiFact, Snopes, TruthOrFiction, etc. and certain health
organizations. The authors proposed a semi-supervised DL model based on Neural Attention
model called Cross-SEAN which leverages huge unlabelled data to improve its performance
and obtained F1-score of 0.95 on CTF dataset.
      </p>
      <p>
        Random Forest (RF), k-Nearest Neighbor (kNN), Logistic Regression (LR) and Support Vector
Machine (SVM), with TF and TF-IDF as features and also DL models, namely, Convolutional
Neural Network (CNN), Long Short Term Memory (LSTM), and Gated Recurrent Network (GRU),
utilizing Glove word embedding have been surveyed by Jiang et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] for fake news detection.
Further, the authors proposed a stacking approach using the above mentioned classifiers and
evaluated that model on two fake news datasets namely, ISOT9 and KDnugget [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and obtained
accuracies of 99.94% and 96.05% respectively. However, in terms of individual performance of
each estimator, LR classifier with an accuracy of 92.82% and RF with an accuracy of 99.87% both
fed with TF-IDF as features outperformed other individual classifiers for both the datasets.
      </p>
      <p>
        Despite the lack of availability of annotated dataset for fake news detection in low resource
languages, Forum for Information Retrieval Evaluation (FIRE) 202010 called for UrduFake11, a
shared task on fake news detection in Urdu language and provided a dataset consisting of 500
real and 400 fake news articles in Urdu [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Balouchzahi et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] proposed various models
based on ML, DL, TL, and hybrid approaches for UrduFake. ML model is a Voting Classifier (VC)
with three estimators namely, Multinomial Naïve Bayes (MNB), Multilayer Perceptron (MLP),
      </p>
      <sec id="sec-2-1">
        <title>9https://www.uvic.ca/engineering/ece/isot/datasets/fake-news/index.php 10http://fire.irsi.res.in/fire/2020/home 11https://www.urdufake2020.cicling.org/home</title>
        <p>and LR, had been fed with set of char and word ngrams features. While DL model has been
developed based on a multi-channel CNN and Skipgram word embedding, an implementation of
Universal Language Model Fine-Tuning (ULMFiT) is based on TL approach. Further, all models
are ensembled as an hybrid model based on majority voting. ML VC outperformed the rest of
the models with an average F1-score of 0.7894.</p>
        <p>
          Tools and modules for the analysis of fake news on social media are not only bounded to
detect the category of a given text but also to identify the spreaders of such false information to
ifnd out the intention beyond sharing fake news. In this direction, PAN at Conference and Labs
of the Evaluation Forum (CLEF) 2020 had called for shared task to identify the spreader of fake
news in Spanish and English. Dataset provided by task organizers consists of 100 tweets per
user (news spreaders) for 300 users. To tackle this task Shashirekha et al. [
          <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
          ] proposed two
classifiers namely, ULMFiT based on TL approach and ensemble of ML classifiers as a voting
classifier. The ULMFiT model initially uses unlabelled data collection from Wikipedia to train
the universal LMs for both English and Spanish languages. As texts from Wikipedia are from
general domain, there is every chance of missing valuable words and features related to fake
news. Hence, the LMs obtained by training on Wikipedia are fine-tuned using the training
data provided by PAN to make LMs more specific for the task of fake news spreader detection.
The Fast.ai12 library is used to develop LMs and classifiers. The proposed ML VC is built with
three estimators, namely, two linear SVM and LR classifiers. After preprocessing the texts by
removing stopwords and punctuation, converting emoji to text and performing lemmatization,
features are extracted. Unigram TFIDF, N-gram TF combined with Doc2vec features are scaled
using MaxAbsScaler and used to train the VC. Chi-square test, Mutual Information, and F-test
are used to select the important features. The final results reported by PAN illustrate that
ULMFiT and ML VC obtained average accuracies of 0.63 and 0.70 respectively.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>
        The performance and efectiveness of ensembling models based on majority voting of ML
classifiers have already been proved for many tasks. Inspired by Gundapu S et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], ensembling
ifne tuned transformer-based models are experimented in this work for fake news detection and
domain identification. Transformer-based models generally include two steps of pre-training
LM and then fine-tuning the model for the new task. Usually as pre-trained models are publicly
available at HuggingFace it is only required to fine-tune LM with a labeled dataset for the target
task. The structure of the proposed model consists of the following four steps:
      </p>
      <sec id="sec-3-1">
        <title>1. Pre-training: pre-trained models available at HuggingFace are used;</title>
        <p>2. Fine-tuning LMs with texts from training set to make LM more domain specific for the
intended task;
3. Training the models obtained in step 2 for classification (each model separately);
4. Ensembling for prediction of labels based on majority voting.</p>
        <p>
          As domain of texts used for pre-training a LM might be diferent from the training set, initially
the individual LMs for transformer-based models namely, BERT [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], Distilbert [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], and
Roberta [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] are fine-tuned using the training data and then these fine-tuned models are used
as estimators in a VC. The main objectives of each estimator are given below:
• BERT is a pre-trained model that employs Masked Language Modeling (MLM) in a
selfsupervised fashion. In other words, only raw texts are used to pre-train the model without
manual annotation. MLM concept along with Next Sentence Prediction (NSP) enables the
model to learn deep representation of a language to extract more eficient features from
texts for downstream tasks.
• Distilbert is a lightweight BERT model. It follows the objectives of BERT model with
distillation loss and returns the same probabilities as BERT and it also utilizes Cosine
embedding concept to generate hidden states as close as BERT model.
• Roberta is an optimized BERT model re-trained with improved training methodology,
more data and hardware resources13. Roberta without NSP concept is similar to BERT
and employs dynamic masking results in changing the masked token during the training
epochs.
        </p>
        <p>Fast-bert14 library with configuration given in Table 2 is used to fine-tune the LMs in (step
2). Except for model and type, the same configuration is used for all the three LMs and
finetuned only for 5 epochs due to resource (RAM and GPU) constraint. Fast-bert associated with
HuggingFace enables a very simple manner to train and evaluate the transformer-based models.
Each individual transformer-based model is trained for 20 epochs.</p>
        <p>As per the general idea of the proposed model illustrated in Figure 2, preprocessing the texts
has been skipped as the models performed better without preprocessing. The same architecture
is followed to construct models for both the subtasks.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <p>
        For any supervised ML task such as TC, annotated data sets are essential to train the model and
enable machine to quickly and clearly understand the input patterns [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Therefore, datasets
provided by the CheckThat! Lab are used and the data collections steps are detailed by Shahi et
al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>Due to mistakes while generating prediction files for the task, the results of the proposed
models reported in the leaderboard are much less than the actual performances of the systems.
13https://towardsdatascience.com/bert-roberta-distilbert-xlnet-which-one-to-use-3d5ab82ba5f8
14https://github.com/utterworks/fast-bert
Therefore, along with the results reported by task organizers, the non-oficial results, i.e., the
actual performances of system evaluated through evaluation script15 provided by the organizer
on re-generated prediction files are also included in this paper. The re-generated and re-evaluated
submissions for both the subtasks can be found in our GitHub page16.</p>
      <p>The training set for Subtask 3A consists of 900 texts in four categories namely, ‘False’, ‘Partially
False’, ‘True’, and ‘Other’. Description of labels is given in Table 1 and label distributions over
training sets are given in Figure 3a. It can be observed that the dataset is highly imbalanced
and as expected this has afected the performance of the proposed model in a negative way. As
per the results reported in leaderboard, for the test set consisting of 364 texts, the proposed
model obtained an F1-score of 0.2334 which is far lower than the expectation due to earlier
mentioned reason of mistakes in prediction files. However, non-oficially the model obtained
0.5034 F1-score on re-generated prediction files.</p>
      <p>A subset of fake news from the dataset of Subtask 3A is used for Subtask 3B. The dataset for
Subtask 3B which includes 318 texts distributed into 6 categories is also imbalanced and the
distribution of labels is shown in Figure 3b. Based on results reported in leaderboard, for the
test set consisting of 137 texts, the proposed model obtained an F1-score of 0.1450 which is far
less than the expectation but actual performance obtained non-oficially is 0.8550 F1-score on
re-generated prediction files.</p>
      <p>Comparison of results both oficial (in leaderboard) and non-oficial, between the proposed
models and top models in the shared task are given in Table 3. It is very much clear that there is
a huge gap between the results reported oficially and the results obtained non-oficially.
Considering actual performances of the proposed model once again the efectiveness of ensembling
classifiers to utilize the strength of single models has been proved.</p>
      <p>15https://gitlab.com/checkthat_lab/clef2021-checkthat-lab/-/tree/master/task3/evaluation
16https://github.com/fazlfrs/CheckThat-_Task3_Submissions</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>In this paper, we, team, MUCIC, have presented the description of the proposed Ensemble of
transformer-based VC models for Fake News Detection, a shared task (Task 3) in
CLEF-2021CheckThat! Lab. Three transformer-based models namely, Roberta, Distilbert, and BERT are
double fine-tuned (once on respective LM and then down streamed for respective TC task)
and ensembled as VC that predicts the label of a given text by majority voting. The proposed
models achieved low F1-scores of 0.233 and 0.145 for Subtask 3A: Multi-Class Fake News
Detection of News Articles and Subtask 3B: Topical Domain Classification of News Articles
respectively, against our expectations due to our mistakes in submission files. However, the
actual performances of the systems show very competitive F1-scores of 0.5309 and 0.8550
for Subtask 3A and Subtask 3B respectively, on re-generated prediction files. Improving the
performances of the proposed models by addressing the problems followed by exploring ML
and DL approaches with various feature sets will be the future work.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgment</title>
      <p>Team MUCIC sincerely appreciates the eforts, guidance and support of the shared task
organizers and reviewers for the valuable comments and suggestions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Balouchzahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          ,
          <article-title>Las for HASOC - learning approaches for hate speech and ofensive content identification</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>145</fpage>
          -
          <lpage>151</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2826</volume>
          /
          <fpage>T2</fpage>
          -6.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Balouchzahi</surname>
          </string-name>
          ,
          <article-title>Ulmfit for twitter fake news spreader profiling</article-title>
          , in: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Eickhof</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Névéol (Eds.), Working Notes of CLEF 2020 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , Thessaloniki, Greece,
          <source>September 22-25</source>
          ,
          <year>2020</year>
          , volume
          <volume>2696</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2696</volume>
          /paper_126.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Anusha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Prakash</surname>
          </string-name>
          ,
          <article-title>Ensemble model for profiling fake news spreaders on twitter</article-title>
          , in: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Eickhof</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Névéol (Eds.), Working Notes of CLEF 2020 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , Thessaloniki, Greece,
          <source>September 22-25</source>
          ,
          <year>2020</year>
          , volume
          <volume>2696</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2696</volume>
          /paper_136.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alawad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gounley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Schaeferkoetter</surname>
          </string-name>
          , H.
          <article-title>-</article-title>
          <string-name>
            <surname>J. Yoon</surname>
          </string-name>
          , X.
          <string-name>
            <surname>-C. Wu</surname>
            ,
            <given-names>E. B.</given-names>
          </string-name>
          <string-name>
            <surname>Durbin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Doherty</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Stroup</surname>
          </string-name>
          , et al.,
          <article-title>Limitations of transformers on clinical text classification</article-title>
          ,
          <source>IEEE journal of biomedical and health informatics</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Prabhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <article-title>Misra, Multi-class text classification using bert-based active learning</article-title>
          ,
          <source>arXiv preprint arXiv:2104.14289</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D. S.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          , T. Mandl,
          <article-title>The CLEF-2021 checkthat! lab on detecting check-worthy claims, previously fact-checked claims, and fake news</article-title>
          , in: D.
          <string-name>
            <surname>Hiemstra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Moens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Perego</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Sebastiani</surname>
          </string-name>
          (Eds.),
          <source>Advances in Information Retrieval - 43rd European Conference on IR Research</source>
          , ECIR
          <year>2021</year>
          ,
          <string-name>
            <surname>Virtual</surname>
            <given-names>Event</given-names>
          </string-name>
          ,
          <year>March</year>
          28 - April 1,
          <year>2021</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , volume
          <volume>12657</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2021</year>
          , pp.
          <fpage>639</fpage>
          -
          <lpage>649</lpage>
          . URL: https: //doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -72240-1_
          <fpage>75</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -72240-1\_
          <fpage>75</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gundapu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mamid</surname>
          </string-name>
          ,
          <article-title>Transformer based automatic covid-19 fake news detection system</article-title>
          ,
          <source>arXiv preprint arXiv:2101.00180</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. PYKL</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Guptha</surname>
          </string-name>
          , G. Kumari,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ekbal</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <article-title>Fighting an infodemic: Covid-19 fake news dataset</article-title>
          , arXiv preprint arXiv:
          <year>2011</year>
          .
          <volume>03327</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          , D. Nandini, FakeCovid
          <article-title>- a multilingual cross-domain fact check news dataset for covid-19</article-title>
          , in: Workshop Proceedings of the 14th
          <source>International AAAI Conference on Web and Social Media</source>
          ,
          <year>2020</year>
          . URL: http://workshop-proceedings.icwsm.org/pdf/2020_14.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W. S.</given-names>
            <surname>Paka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kaushik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sengupta</surname>
          </string-name>
          , T. Chakraborty,
          <article-title>Cross-sean: A cross-stitch semi-supervised neural attention model for covid-19 fake news detection</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>107</volume>
          (
          <year>2021</year>
          )
          <fpage>107393</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. U.</given-names>
            <surname>Haq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Saboor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <article-title>A novel stacking approach for accurate detection of fake news</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2021</year>
          )
          <fpage>22626</fpage>
          -
          <lpage>22639</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P. H. A.</given-names>
            <surname>Faustini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. F.</given-names>
            <surname>Covões</surname>
          </string-name>
          ,
          <article-title>Fake news detection in multiple platforms and languages</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>158</volume>
          (
          <year>2020</year>
          )
          <fpage>113503</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Amjad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sidorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zhila</surname>
          </string-name>
          ,
          <article-title>Data augmentation using machine translation for fake news detection in the urdu language</article-title>
          ,
          <source>in: Proceedings of The 12th Language Resources and Evaluation Conference</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>2537</fpage>
          -
          <lpage>2542</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Balouchzahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          ,
          <article-title>Learning models for urdu fake news detection</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>474</fpage>
          -
          <lpage>479</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2826</volume>
          /
          <fpage>T3</fpage>
          -7.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          , T. Wolf,
          <article-title>Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter</article-title>
          , arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>01108</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Amused:</surname>
          </string-name>
          <article-title>An annotation framework of multi-modal social media data</article-title>
          , arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>00502</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          , T. Mandl,
          <article-title>Overview of the CLEF-2021 CheckThat! lab task 3 on fake news detection</article-title>
          , in: Working Notes of CLEF 2021-
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , CLEF '
          <year>2021</year>
          , Bucharest, Romania (online),
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>