<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Check square at CheckThat! 2020: Claim Detection in Social Media via Fusion of Transformer and Syntactic Features</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gullal S. Cheema</string-name>
          <email>gullal.cheema@tib.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sherzod Hakimov</string-name>
          <email>sherzod.hakimov@tib.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ralph Ewerth</string-name>
          <email>ralph.ewerth@tib.eu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>L3S Research Center, Leibniz University Hannover</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>TIB Leibniz Information Centre for Science and Technology</institution>
          ,
          <addr-line>Hannover</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this digital age of news consumption, a news reader has the ability to react, express and share opinions with others in a highly interactive and fast manner. As a consequence, fake news has made its way into our daily life because of very limited capacity to verify news on the Internet by large companies as well as individuals. In this paper, we focus on solving two problems which are part of the factchecking ecosystem that can help to automate fact-checking of claims in an ever increasing stream of content on social media. For the rst problem, claim check-worthiness prediction, we explore the fusion of syntactic features and deep transformer Bidirectional Encoder Representations from Transformers (BERT) embeddings, to classify check-worthiness of a tweet, i.e. whether it includes a claim or not. We conduct a detailed feature analysis and present our best performing models for English and Arabic tweets. For the second problem, claim retrieval, we explore the pre-trained embeddings from a Siamese network transformer model (sentence-transformers) speci cally trained for semantic textual similarity, and perform KD-search to retrieve veri ed claims with respect to a query tweet.</p>
      </abstract>
      <kwd-group>
        <kwd>Check-Worthiness</kwd>
        <kwd>Fact-Checking</kwd>
        <kwd>Social Media</kwd>
        <kwd>Twitter</kwd>
        <kwd>COVID19</kwd>
        <kwd>SVM</kwd>
        <kwd>BERT</kwd>
        <kwd>Retrieval</kwd>
        <kwd>Text Classi cation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Social media is increasingly becoming the main source of news for so many
people. With around 2.5 billion Internet users, 12% receive breaking news from
Twitter instead of traditional media according to a 2018 survey report [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ].
Fake news in general can be de ned [
        <xref ref-type="bibr" rid="ref52">52</xref>
        ] as fabrication and manipulation of
information and facts with the main intention of deceiving the reader. As a result,
fake news can have several undesired and negative consequences. For example,
recent news around COVID-19 pandemic with non-veri ed claims, that masks
lead to rise in carbon dioxide levels caused an online movement to not wear
masks. With ease of access and sharing news on Twitter, any news spreads much
faster from the moment an event occurs in any part of the world. Although, the
survey report [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ] found that almost 60% of users expect news on social media
to be inaccurate, it still leaves millions of people who will spread fake news
expecting it to be true.
      </p>
      <p>Considering the vast amount of news that spreads everyday, there has been
a rise in independent fact-checking projects like Snopes, Alt News, Our.News,
who investigate the news that spread online and publish the results for public
use. Most of these independent projects rely on manual e orts that are time
consuming which makes it harder to keep up with rate of news production.
Therefore, it has become very important to develop tools that can process news
at a rapid rate and provide news consumers with some kind of an authenticity
measure that re ects the correctness of claims in the news.</p>
      <p>
        In this paper, we focus on two sub-problems in CheckThat! 2020 [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ]3 that
are a part of larger fact-checking ecosystem. In the rst task, we focus on
learning a model that can recognize check-worthy claims on Twitter. We present a
solution that works for both English [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ] and Arabic [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] tweets. Some examples
of tweets with claims are classi ed whether it is a check-worthy or not, shown in
Table 1. One can see that the claims which are not check-worthy look like
personal opinions and do not pose any threat to a larger audience. We explore the
fusion of syntactic features and deep transformer Bidirectional Encoder
Representations from Transformers (BERT) embeddings, to classify check-worthiness
of a tweet. We use Part-of-speech (POS) tags, named entities, and dependency
relations as syntactic features and a combination of hidden layers in BERT to
compute tweet embedding. Before learning the model with a Support Vector
Machine (SVM) [
        <xref ref-type="bibr" rid="ref51">51</xref>
        ], we use Principal Component Analysis (PCA) [
        <xref ref-type="bibr" rid="ref57">57</xref>
        ] for
dimensionality reduction. In the second task, we focus on learning a model that
can accurately retrieve veri ed claims w.r.t a query claim, where query claim
is a tweet and veri ed claims are snippets from actual documents. The veri ed
claim is true and thus acts as the evidence or support for the query tweet. Some
example pairs of tweets and claims can be seen in Table 2, which shows that the
pairs share lots of contextual information which makes this task a semantic
textual similarity problem. For this reason, we explore the pre-trained embeddings
from a Siamese network transformer model (sentence-transformers) speci cally
trained for semantic textual similarity and perform KD-search to retrieve claims.
We share the source code for both tasks publicly with the community. 4
      </p>
      <p>The remainder of the paper is organized as follows. Section 2 brie y
discusses about previous works on fake news detection and CheckThat! tasks in
particular. Section 3 presents details of our approach for Task-1 and Task-2.
Section 4 describes the experimental details and results. Section 5 summarizes
our conclusion and future research directions.</p>
    </sec>
    <sec id="sec-2">
      <title>3 https://sites.google.com/view/clef2020-checkthat/ 4 https://github.com/cleopatra-itn/claim_detection</title>
      <p>Tweet
Dear @VP Pence: What are you hiding from the
American people? Why won't you let the people
see and hear what experts are saying about the
#CoronaOutbreak?
Greeting my good friends from the #US the
#Taiwan way. Remember: to better prevent the
spread of #COVID19, say no to a handshake &amp;
yes to this friendly gesture. Check it out:
Corona got these ights cheap as hell I gotta job
interview in Greece Monday
My mum has a PhD on Corona Virus from
WhatsApp University
This is why the beaches haven't closed in
Florida, and why they've had minimal COVID-19
prevention. Absolute dysfunction. &lt;link&gt;
COVID-19 cases in the Philippines jumped
from 24 to 35 in less than 12 hours. This is
seriously ALARMING. Stay safe everyone!
&lt;link&gt;
0
0
1
1
1
1
0
0
0
0
1
1</p>
      <sec id="sec-2-1">
        <title>Related Work</title>
        <p>
          Fake news has been studied from di erent perspectives in the last ve years, like
factuality or credibility detection [
          <xref ref-type="bibr" rid="ref15 ref20 ref21 ref38 ref42">38,15,42,21,20</xref>
          ], rumour detection [
          <xref ref-type="bibr" rid="ref47 ref59 ref60 ref61">61,60,47,59</xref>
          ],
propagation in networks [
          <xref ref-type="bibr" rid="ref28 ref32 ref35 ref45">28,32,45,35</xref>
          ], use of multiple modalities [
          <xref ref-type="bibr" rid="ref27 ref48 ref54">27,54,48</xref>
          ] and
also as an ecosystem of smaller sub-problems like in CheckThat! [
          <xref ref-type="bibr" rid="ref11 ref2 ref33">33,11,2</xref>
          ]. For
social media in particular, Shu et al. [
          <xref ref-type="bibr" rid="ref46">46</xref>
          ] studied and provided a
comprehensive review of fake news detection with characterizations from psychology and
social science, and existing computational algorithms from data mining
perspective. The fact that Twitter has become a source of news for so many people,
researchers have extensively used the platform to formulate problems, extract
data and test their algorithms. For instance, Zubiaga et. al. [
          <xref ref-type="bibr" rid="ref61">61</xref>
          ] extracted tweets
around breaking news and used Conditional Random Fields to exploit context
during the sequential learning process for rumour detection. Buntain et. al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]
studied three large Twitter datasets and developed models to predict accuracy
assessments of fake news by crowd-sourced workers and journalists. While many
approaches rely on tweet content for detecting fake news, there has been a rise
in methods that exploit user characteristics and metadata to model the problem
as fake news propagation. For example, Liu et. al. [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] modeled the propagation
path of each news story as a multivariate time series over users who engaged in
spreading the news via tweets. They further classi ed the fake news using Gated
Recurrent Unit (GRU) and Convolutional Neural Networks (CNN) to capture
the global and local variations of user characteristics respectively. Monti et. al.
[
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] went a step further and used a hybrid feature set including user
characteristics, social network structure and tweet content. They modeled the problem as
binary prediction using a Graph CNN resulting in a highly accurate fake news
detector.
        </p>
        <p>
          Besides fake news detection, a sub task to predict check-worthiness of claims
has also been explored recently mostly in political context. For example, Hassan
et. al. [
          <xref ref-type="bibr" rid="ref21 ref22">21,22</xref>
          ] proposed a system that predicts the check-worthiness of a
statement made by presidential candidates using SVM [
          <xref ref-type="bibr" rid="ref51">51</xref>
          ] classi er and combination
of lexical and syntactic features. They also compared their results with
factchecking organizations like CNN5 and PolitiFact6. Later, in CheckThat! 2018
[
          <xref ref-type="bibr" rid="ref33">33</xref>
          ], several methods were proposed to improve the check-worthiness of claims
in political debates. Best methods used a combination of lexical and syntactic
features like Bag of Words (BoW), Parts-of-Speech (POS) tags, named Entities,
sentiment, topic modeling, dependency parse trees and word embeddings [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ].
Various classi ers were built using either Recurrent Neural Networks (RNN)
[
          <xref ref-type="bibr" rid="ref16 ref62">16,62</xref>
          ], gradient boosting [
          <xref ref-type="bibr" rid="ref58">58</xref>
          ], k-nearest neighbor [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] or SVM [
          <xref ref-type="bibr" rid="ref62">62</xref>
          ]. In 2019
edition of CheckThat! [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], in addition to using lexical and syntactic features
[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], top approaches relied on learning richer content embeddings and utilized
external data for better performance. For example, Hansen et. al. [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] used word
embeddings and syntactic dependency features as input to an LSTM network,
5 http://www.cnn.com
6 https://www.politifact.com/
enriched the dataset with additional samples from Claimbuster system [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] and
trained the network with a contrastive ranking loss. Favano et. al. [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] trained
a neural network with Standard Universal Sentence Encoder (SUSE) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]
embeddings of the current sentence and previous two sentences as context. Another
approach by Su et. al. [
          <xref ref-type="bibr" rid="ref50">50</xref>
          ] used co-reference resolution to replace pronouns with
named entities to get a feature representation with bag of words, named
entity similarity and relatedness. Other than political debates, Jaradat et. al. [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]
proposed an online multilingual check-worthiness system that works for di
erent sources (debates, news articles, interviews) in English and Arabic . They
use actual annotated data from reputable fact-checking organizations and use
best performing feature representations from previous approaches. For tweets in
particular, Majithia et. al. [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ] proposed a system to monitor, search and
analyze factual claims in political tweets with Claimbuster [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] at the backend for
check-worthiness. Lastly, Dogan et. al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] also conducted a detailed study on
detecting check-worthy tweets in U.S. politics and proposed a real-time system
to lter them.
3
3.1
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Proposed Approach</title>
        <sec id="sec-2-2-1">
          <title>Task-1: Tweet Check-Worthiness Prediction</title>
          <p>
            Check-Worthiness prediction is the task of predicting whether a tweet includes
a claim that is of interest to a large audience. Our approach is motivated by
the successful use of lexical, syntactic and contextual features in the previous
editions of CheckThat! check-worthiness task for political debates. Given that
this task contains less amount of training data, we approached this problem
with the idea of creating a rich feature representation, reducing the dimensions
of large feature set with PCA [
            <xref ref-type="bibr" rid="ref57">57</xref>
            ] and then learning the model with a SVM. In
doing so, our goal is also to understand which features are the most important for
check-worthiness prediction from tweet content. As context is very important for
downstream NLP tasks, we experiment with word embeddings (word2vec [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ],
GloVe [
            <xref ref-type="bibr" rid="ref37">37</xref>
            ]) and BERT [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ] embeddings to create a sentence representation of each
tweet. Our pre-processing and feature extraction is agnostic to the topic of the
tweet so that it can be applied to any domain. Next, we provide details about
all the features used, their extraction and the encoding process. Our overall
approach can be seen in Figure 2.
          </p>
          <p>
            Pre-processing We use two publicly available pre-processing tools for English
and Arabic tweets. We use Baziotis et. al.'s [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ] tool for English to apply the
following normalization steps: tokenization, lower-casing, removal of punctuation,
spell correction, normalize hashtags, all-caps, censored, elongated and repeated
words, and terms like URL, email, phone, user mentions. We use Stanford Stanza
[
            <xref ref-type="bibr" rid="ref39">39</xref>
            ] toolkit to pre-process Arabic tweets by applying the following normalization
steps: tokenization, multi-word token expansion and lemmatization.
          </p>
          <p>
            In the case of extracting word embeddings from a transformer network, we
use the raw text as the networks have their own tokenization process.
Syntactic Features We use the following syntactic features for English and
Arabic tasks: Parts-of-Speech (POS) tags, named entities (NE) and dependency
parse tree relations. We use the pre-processed text and run o -the-shelf tools to
extract syntactic information of tweets and then convert each group of
information to feature sets. For English we used spaCy[
            <xref ref-type="bibr" rid="ref24">24</xref>
            ] and Stanford Stanza [
            <xref ref-type="bibr" rid="ref39">39</xref>
            ] for
Arabic tweets to extract the following syntactic features. In all the features, we
experiment with keeping and removing stop-words to evaluate their a ect.
          </p>
          <p>Part-of-Speech: For both English and Arabic, we extract 16 POS tags in
total and through our empirical evaluation we nd that the following eight tags to
be the most useful when used as features: NOUN, VERB, PROPN, ADJ, ADV,
NUM, ADP, PRON. For Arabic, the additional four tags are useful features:
DET, INTJ, AUX, PART. We used the chosen set of POS tags for respective
language to encode the syntactic information of tweets.</p>
          <p>Named Entities: We identi ed the following named entity types to be the
most important features through our evaluation: (GPE, PERSON, ORG, NORP,
LOC, DATE, CARDINAL, TIME, ORDINAL, FAC, MONEY) for English and
(LOC, PER, ORG, MISC) for Arabic. We also found that while developing
feature combinations named entities do not add much value to overall accuracy,
and hence our primary and contrastive submissions do not include them.</p>
          <p>Syntactic Dependencies: these features are constructed using dependency
relation between tokens in a given tweet. We use the dependency relation between
two nodes in the parsed tree if the the child and parent nodes' POS tags are one
of the following ADJ, ADV, NOUN, PROPN, VERB or NUM. All dependency
relations that match the de ned constraint are converted into the triplet relation
such as (child node-POS, dependency-relation, parent-POS ) and pairs such as
(child node-POS, dependency-relation) where the relation is not part of a feature
representation. This process is shown in Figure 1. We found that the features
based on pairs of child and parent node perform better than the triplet feature.
The dimension of the feature vector for English and Arabic is 133 and 134
respectively.</p>
          <p>For encoding a feature, we get a histogram vector which contains the number
of type of tag, named entity or syntactic relation pair. The process of feature
encoding is shown in Figure 1. Finally, we normalize each type of feature with
maximum value in the vector.</p>
          <p>
            Average Word Embeddings One simple way to get a contextual
representation of a sentence is to average the word embeddings of each token in a given
sentence. For this purpose, we experiment with three types of word embeddings
pre-trained on three di erent sources for English: GloVe embeddings [
            <xref ref-type="bibr" rid="ref37">37</xref>
            ] trained
on Twitter and Wikipedia, word2vec embeddings [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ] trained on Google News,
and FastText [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ] embeddings trained on multiple sources. In addition, we also
experiment with removing stop-words from the average word representation, as
stop-words can dominate in the average and result in less meaningful sentence
representation. For Arabic, we use word2vec embeddings that are trained on
Arabic tweets and Arabic Wikipedia [
            <xref ref-type="bibr" rid="ref49">49</xref>
            ].
          </p>
          <p>
            Transformer Features Another way to extract contextual features is to use
BERT [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ] embeddings that are trained using the context of the word in a
sentence. BERT is usually trained on a very large text corpus which makes them
very useful for o -the-shelf feature extraction and ne-tuning for downstream
tasks in NLP. To get one embedding per tweet, we follow the observations made
in [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ] that, di erent layers of BERT capture di erent kinds of information, so
an appropriate pooling strategy should be applied depending on the task. The
paper also suggests that the last four hidden layers of the network are good for
transfer learning tasks and thus we experiment with 4 di erent combinations,
i.e. concatenate last 4 hidden layers, average of last 4 hidden layers, last hidden
layer and 2nd last hidden layer. We normalize the nal embedding so that l2
norm of the vector is 1. We also experimented with BERT's pooled sentence
embedding that is encoded in the CLS (class) tag, which performed signi cantly
poorer than the pooling strategies we employed. For Arabic, we only
experimented with a sentence-transformer [
            <xref ref-type="bibr" rid="ref40">40</xref>
            ] that is trained on multilingual training
corpus and outputs a sentence embedding for each tweet/sentence.
          </p>
          <p>Sentence Representation: To get the overall representation of the tweet,
we concatenate all the syntactic features together with either average word
embedding or BERT-based transformer features and then apply PCA for
dimensionality reduction. SVM classi er is trained on the feature vectors of tweets to
output a binary decision (check worthy or not). The overall process is shown in
Figure 2.</p>
          <p>Fig. 2: Proposed Approach for Check-Worthiness Prediction
3.2</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>Task-2: Claim Retrieval</title>
          <p>Claim Retrieval is the task of retrieving the most similar already veri ed claim
to the query claim. For this task, it is important that the feature representation
captures the meaning and context of words and phrases so that query matches
the correct veri ed claim. Therefore, we relied on a triplet-network setting, where
the network could be trained with triplets consisting of an anchor sample a,
positive sample p and a negative sample n. We use triplet loss to ne-tune a
pre-trained sentence embedding network, such that the distance between a and
p is smaller than the distance between a and n using the following loss function.</p>
          <p>Loss =</p>
          <p>N
X[jjSi</p>
          <p>a
n=1</p>
          <p>p 2
Si jj2</p>
          <p>a
jjSi</p>
          <p>Sinjj22 + m]+
(1)
where Sia, Sip and Sin are triplet sentence embeddings and m is the margin
(set to 1), N is the number of samples in the batch.</p>
          <p>
            As each veri ed claim is a tuple consisting of text and title, we create two
triplets for every true tweet-claim pair, i.e., (anchor tweet, true claim text,
negative claim text) and (anchor tweet, true claim title, negative claim title). This
increases the number of positive samples for training as there are only 800
samples and one true claim for every tweet. To get negative claims, we select 3 claims
with highest cosine similarity that are not the true claims for the anchor tweet
using the pre-trained sentence-transformer embeddings. For pre-processing, we
use Baziotis et. al.'s [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ] tool for processing the tweets to remove URL, email,
phone, user mentions, as the claim text or title do not contain any such
information.
          </p>
          <p>
            As retrieval is a search task, we used KD-Tree search to nd the most similar
already veri ed claim that has the minimum Euclidean distance to the query
claim. The sentence embeddings extracted from the network are used to build a
KD-Tree and for each query claim, top 1000 veri ed claims are extracted from
the tree for evaluation. For building the KD-Tree, we average the sentence
embeddings of claim text and claim title, as it performs better than just using
either claim or title. In our ablation study, we directly compute the cosine
similarity between each query tweet and all the veri ed claims, and pick the top
1000 (highest cosine similarity) veri ed claims for evaluation. We conduct the
second evaluation because building a KD-Tree can a ect the retrieval accuracy.
Sentence Transformers for Textual Similarity As a backbone network to
extract sentence embeddings and ne-tuning with triplet loss, we use the recently
proposed Sentence-BERT [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ] that learns the embeddings in a Siamese (pairs) and
triplet network settings. We experiment with the pre-trained Siamese Network
models trained on SNLI (Stanford Natural Language Inference) [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] and STSb
(Semantic Textual Similarity benchmark) [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] datasets that have been shown to
perform very well for semantic textual similarity.
4
4.1
          </p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Experiments and Results</title>
        <sec id="sec-2-3-1">
          <title>Task-1: Tweet Check-Worthiness Prediction</title>
          <p>Dataset and Training Details English dataset consists of training,
development (dev) and test splits with 672, 150 and 140 tweets respectively on the
topic of COVID-19. We perform grid search using development set to nd the
best parameters. Arabic dataset consists of training and test splits with 1500
tweets on 3 topics and 6000 tweets on 12 topics respectively with 500 tweets on
each topic. For validation purpose, we keep 10% (150 samples) from the training
data as development set. The o cial ranking of submitted system for this task is
based on Mean Average Precision (MAP) and Precision@30 (P@30) for English
and Arabic datasets, respectively.</p>
          <p>
            To train the SVM models for both English and Arabic, we perform grid
search over PCA energy (%) conservation, regularization parameter C and RBF
kernel's gamma. Parameters range for PCA varies from 100% (original features)
to 95% with decrements of 1, and both C and gamma vary between -3 to 3
on a log-scale with 30 steps. For faster training on a large grid search, we use
ThunderSVM [
            <xref ref-type="bibr" rid="ref55">55</xref>
            ] which takes advantage of a GPU or a multi-core system to
speed up SVM training.
Results Our submissions used the best models that we obtained from the grid
search and are brie y discussed below.
          </p>
          <p>
            English: We made 3 submissions in total. Our primary (Run-1) and 2nd
contrastive (Run-3) submission uses sentence embeddings computed from
BERTlarge word embeddings as discussed in the proposed work section. In addition,
both submissions use POS tag and dependency relation features. Interestingly,
we found that the best performing sentence embeddings did not include
stopwords. The primary submission (Run-1) uses an ensemble of predictions from
three models trained on concatenated last 4 hidden layers, average of last 4
hidden layers and 2nd last hidden layer. The 2nd contrastive submission (Run-3)
uses predictions from the model trained on the best performing sentence
embedding computed from concatenating last 4 hidden layers. Our 1st contrastive
submission (Run-2) uses an ensemble of predictions from three models trained
with GloVe[
            <xref ref-type="bibr" rid="ref37">37</xref>
            ] on Twitter with 25, 50 and 100-dimensional embeddings but with
the same POS tag and dependency relation features. We use majority voting to
get the nal prediction and mean of decision values to get the nal decision value.
We found that removing the stop-words to compute average of word embeddings
actually degraded the performance and hence included them in the average.
          </p>
          <p>
            We also add some additional results to see the e ect of stop-words, POS
tags, named entities, dependency relations and ensemble predictions in Table 3.
The e ect of stop-words can be clearly seen in alternative runs of Run-1 and
Run-3, where the MAP clearly drops by 1-2 points. Similarly, the negative e ect
of removing POS tag and dependency relation features can be seen in rest of the
alternative runs. Lastly, adding named entity features to the original submissions
also decreases the precision by 1-2 points. This might be because the tweets have
very few named entities and are not useful to distinguish between check-worthy
and not check-worthy claims. For comparison with other teams in the challenge,
we show top 3 results at the bottom of the table for reference. Team Accenture
[
            <xref ref-type="bibr" rid="ref56">56</xref>
            ] ne-tuned a RoBERTa model with an extra mean pooling and a dropout
layer to prevent over tting. Team Alex [
            <xref ref-type="bibr" rid="ref34">34</xref>
            ] experimented with di erent tweet
pre-processing techniques and various transformer models together with
logistic regression and SVM. Their main submission used logistic regression trained
on 5-fold predictions from RoBERTa concatenated with tweet metadata. Team
QMUL-SDS [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] ne-tuned a BERT model pre-trained speci cally on COVID
twitter data.
          </p>
          <p>
            Arabic There are a total of four submissions that we made in this task.
Our best performing submission (Run-1) uses 100-dimensional word2vec Arabic
embeddings trained on a Twitter corpus [
            <xref ref-type="bibr" rid="ref49">49</xref>
            ] in combination with POS tag
features. Our second and third submissions are redundant in terms of feature use,
so we only mention the second one (Run-2) here. In addition features used in
rst submission, it uses dependency relation features and 300-dimensional
Twitter embeddings instead of 100-dimensional. Our last submission (Run-3) uses
only pre-trained multilingual sentence-transformer7 [
            <xref ref-type="bibr" rid="ref41">41</xref>
            ] that is trained on 10
languages including Arabic. In the rst three submissions, we removed the
stop
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>7 https://github.com/UKPLab/sentence-transformers</title>
      <p>
        Run
Run-1
Run-2
Run-3
Run-1-1
Run-1-2
Run-1-3
Run-1-4
Run-3-1
Run-3-2
Run-3-3
Run-3-4
[
        <xref ref-type="bibr" rid="ref56">56</xref>
        ]
[
        <xref ref-type="bibr" rid="ref34">34</xref>
        ]
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
      </p>
      <p>X
X
X</p>
      <p>
        X
X
X
X
X
X
words from all the features as keeping them resulted in a poorer performance.
Precision@K and Average Precision (AP) results on the test set are shown in the
same order in Table 4. O cial metric for ranking is P@30. For comparison with
other teams in the challenge,we show top 3 results at the bottom of the table for
reference. Team Accenture [
        <xref ref-type="bibr" rid="ref56">56</xref>
        ] experimented with and ne-tuned three di erent
pre-trained Arabic BERT models and used external data to increase the positive
instances. Team TOBB-ETU [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] used logistic regression and experimented with
Arabic BERT and word embeddings together to classify tweets. Team UB ET
[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] used a multilingual BERT for ranking tweets by check-worthiness.
4.2
      </p>
      <sec id="sec-3-1">
        <title>Task-2: Claim Retrieval</title>
        <p>Dataset and Training Details The dataset in this task has 1,003 tweets for
training and 200 tweets for testing. These tweets are to be matched against a set
10,373 veri ed claims. From the training set, 197 tweets are kept for validation.
To ne-tune the sentence-transformer network with the triplet loss, we use a
batch size of eight and train the network for two epochs. The o cial ranking of
this is based on Mean Average Precision@5 (MAP@5). All tweets and veri ed
claims are in English.</p>
        <p>X
X
X
X
X
X
X
X
X</p>
        <p>BERT
GloVe
BERT
BERT
BERT
BERT
BERT
BERT
BERT
BERT
BERT</p>
        <p>MAP</p>
        <p>
          Our primary (Run-1) and 2nd contrastive (Run-3) submission uses
BERTbase and BERT-large pre-trained on SNLI dataset with sentence embedding
pooled from the CLS and MAX tokens respectively. We ne-tune these two
networks with the triplet loss. On the contrary, our 1st contrastive submission
(Run-2) uses multilingual DistilBERT model [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ] trained on 10 languages
including English. This model is directly used to test the pre-trained embeddings.
Results Interestingly, pre-trained embeddings extracted from multilingual
DistilBERT without any ne-tuning turn out to be better for semantic similarity
than ne-tuned monolingual BERT models. Having said that, the ne-tuned
monolingual BERT models do perform better than extracted pre-trained
embeddings and the di erence can be seen in Run-1-2 and Run-3-2 in Table 5. We
also try to ne-tune the multilingual model which drastically decreases the
retrieval performance. The decrease can be attributed to the pre-training process
[
          <xref ref-type="bibr" rid="ref41">41</xref>
          ] in which the model was trained in a teacher-student knowledge distillation
learning framework and on multiple languages. As stated in the proposed work
section, we conduct a second evaluation to retrieve the claims with highest
similarity without KD-Search and the results are signi cantly better as shown in
Table 5. For comparison with other teams in the challenge, we have shown top
3 primary submissions at the bottom of the table for reference. Team Buster.AI
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] investigated sentence similarity using transformer models, and experimented
with multimodality and data augmentation. Team UNIPI-NLE [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ] ne-tuned
a sentence-BERT in two steps, rst to predict the cosine similarity of positive
and negative pairs, followed by a binary classi cation of whether a tweet-claim
pair is a correct match or not. Team UB ET [
          <xref ref-type="bibr" rid="ref53">53</xref>
          ] experimented with three
different models to rank the veri ed claims and their main submission used a DPH
Divergence from Randomness (DFR) term weighting model.
Run
Run-1
Run-2
Run-3
In this paper, we have presented our solutions for two tasks in CLEF
CheckThat! 2020. In the rst task, we used syntactic, contextual features and SVM for
predicting the check-worthiness of tweets in Arabic and English. For syntactic
features, we evaluated Parts-of-Speech tags, named entities and syntactic
dependency relations, and used the best feature sets for both languages. In the case
of contextual features, we evaluated di erent word embeddings, BERT models
and sentence-transformers to capture the semantics of each tweet or sentence.
For future work, we would like to evaluate the possibility of using relevant
metadata and other modalities like images and videos present in tweets for claim's
check-worthiness. In the second task, we evaluated monolingual and
multilingual sentence-transformers to retrieve veri ed claims for the query tweet. We
found that o -the-shelf multilingual sentence-transformer is very well suited for
semantic textual similarity task than other monolingual BERT models.
        </p>
        <sec id="sec-3-1-1">
          <title>Acknowledgements</title>
          <p>This project has received funding from the European Union's Horizon 2020
research and innovation programme under the Marie Sklodowska-Curie grant
agreement no 812997.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Alkhalifa</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yoong</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kochkina</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubiaga</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liakata</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>QMUL-SDS at CheckThat! 2020: Determining COVID-19 tweet check-worthiness using an enhanced CT-BERT with numeric expressions</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Barron-Ceden~o,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            , Da San Martino, G.,
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Hamdan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Sheikh Ali</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          : Overview of CheckThat! 2020:
          <article-title>Automatic identi cation and veri cation of claims in social media</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baziotis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelekis</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doulkeridis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Datastories at semeval
          <article-title>-2017 task 4: Deep lstm with attention for message-level and topic-based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval2017)</source>
          . pp.
          <volume>747</volume>
          {
          <fpage>754</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (
          <year>August 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bouziane</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cluzeau</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mardas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sadeq</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : Buster.AI at CheckThat! 2020:
          <article-title>Insights and recommendations to improve fact-checking.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bowman</surname>
            ,
            <given-names>S.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Angeli</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potts</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.D.:</given-names>
          </string-name>
          <article-title>A large annotated corpus for learning natural language inference</article-title>
          .
          <source>arXiv preprint arXiv:1508.05326</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Buntain</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Golbeck</surname>
          </string-name>
          , J.:
          <article-title>Automatically identifying fake news in popular twitter threads</article-title>
          .
          <source>In: 2017 IEEE International Conference on Smart Cloud (SmartCloud)</source>
          . pp.
          <volume>208</volume>
          {
          <fpage>215</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Cer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Diab</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agirre</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez-Gazpio</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Specia</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation</article-title>
          .
          <source>arXiv preprint arXiv:1708.00055</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Cer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kong</surname>
          </string-name>
          , S.y.,
          <string-name>
            <surname>Hua</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Limtiaco</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>John</surname>
          </string-name>
          , R.S.,
          <string-name>
            <surname>Constant</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guajardo-Cespedes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tar</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , et al.:
          <article-title>Universal sentence encoder</article-title>
          . arXiv preprint arXiv:
          <year>1803</year>
          .
          <volume>11175</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Dogan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , et al.:
          <article-title>Detecting Real-time Check-worthy Factual Claims in Tweets Related to US Politics</article-title>
          .
          <source>Ph.D. thesis</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barron-Cedeno</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasanain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suwaileh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Da San Martino, G.,
          <string-name>
            <surname>Atanasova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Overview of the clef-2019 checkthat! lab: automatic identi cation and veri cation of claims</article-title>
          . In:
          <article-title>International Conference of the Cross-Language Evaluation Forum for European Languages</article-title>
          . pp.
          <volume>301</volume>
          {
          <fpage>321</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Favano</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carman</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lanzi</surname>
            ,
            <given-names>P.L.</given-names>
          </string-name>
          :
          <article-title>Theearthis at's submission to clef'19checkthat! challenge</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Gasior</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Przybyla</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>The ipipan team participation in the check-worthiness task of the clef2019 checkthat! lab (</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Ghanem</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>M.y</article-title>
          .,
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Upv-inaoe-autoritas-check that: Preliminary approach for checking worthiness of claims</article-title>
          .
          <source>CLEF</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Giachanou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Leveraging emotional signals for credibility detection</article-title>
          .
          <source>In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          . pp.
          <volume>877</volume>
          {
          <issue>880</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simonsen</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lioma</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The copenhagen team participation in the check-worthiness task of the competition of automatic identi cation and veri cation of claims in political debates of the clef-2018 checkthat! lab</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simonsen</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lioma</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Neural weakly supervised fact check-worthiness detection with contrastive sampling-based ranking loss</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Hasanain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
          </string-name>
          , T.: bigIR at CheckThat! 2020:
          <article-title>Multilingual BERT for ranking Arabic tweets by check-worthiness</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Hasanain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haouari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suwaileh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ali</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hamdan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , BarronCeden~o,
          <string-name>
            <given-names>A.</given-names>
            , Da San Martino, G.,
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          : Overview of CheckThat! 2020 Arabic:
          <article-title>Automatic identi cation and veri cation of claims in social media</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arslan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tremayne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Toward automated fact-checking: Detecting check-worthy factual claims by claimbuster</article-title>
          .
          <source>In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <year>1803</year>
          {
          <year>1812</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tremayne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Detecting check-worthy factual claims in presidential debates</article-title>
          .
          <source>In: Proceedings of the 24th acm international on conference on information and knowledge management</source>
          . pp.
          <year>1835</year>
          {
          <year>1838</year>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tremayne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arslan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Comparing automated factual claim detection against judgments of journalism organizations</article-title>
          .
          <source>In: Computation+ Journalism Symposium</source>
          . pp.
          <volume>1</volume>
          {
          <issue>5</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , G.,
          <string-name>
            <surname>Arslan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caraballo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jimenez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gawsane</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joseph</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kulkarni</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nayak</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          , et al.:
          <article-title>Claimbuster: the rst-ever end-toend fact-checking system</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>10</volume>
          (
          <issue>12</issue>
          ),
          <year>1945</year>
          {
          <year>1948</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Honnibal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montani</surname>
          </string-name>
          , I.:
          <article-title>spaCy 2: Natural language understanding with bloom embeddings, convolutional neural networks and incremental parsing</article-title>
          . To appear (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Jaradat</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gencheva</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Barron-Ceden~o,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Marquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          : Claimrank:
          <article-title>Detecting check-worthy claims in arabic and english</article-title>
          . arXiv preprint arXiv:
          <year>1804</year>
          .
          <volume>07587</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Kartal</surname>
            ,
            <given-names>Y.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kutlu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : TOBB ETU at CheckThat! 2020:
          <article-title>Prioritizing English and Arabic claims based on check-worthiness</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Khattar</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goud</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varma</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Mvae: Multimodal variational autoencoder for fake news detection</article-title>
          .
          <source>In: The World Wide Web Conference</source>
          . pp.
          <volume>2915</volume>
          {
          <issue>2921</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Y.F.B.</given-names>
          </string-name>
          :
          <article-title>Early detection of fake news on social media through propagation path classi cation with recurrent and convolutional networks</article-title>
          .
          <source>In: ThirtySecond AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Majithia</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arslan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lubal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jimenez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arora</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caraballo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Claimportal: Integrated monitoring, searching, checking, and analytics of factual claims on twitter</article-title>
          .
          <source>In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations</source>
          . pp.
          <volume>153</volume>
          {
          <issue>158</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Puhrsch</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Advances in pretraining distributed word representations</article-title>
          .
          <source>arXiv preprint arXiv:1712.09405</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In: Advances in neural information processing systems</source>
          . pp.
          <volume>3111</volume>
          {
          <issue>3119</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Monti</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frasca</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eynard</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mannion</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bronstein</surname>
            ,
            <given-names>M.M.:</given-names>
          </string-name>
          <article-title>Fake news detection on social media using geometric deep learning</article-title>
          .
          <source>arXiv preprint arXiv:1902</source>
          .
          <volume>06673</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barron-Cedeno</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suwaileh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marquez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaghouani</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Atanasova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kyuchukov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Da San Martino, G.:
          <article-title>Overview of the clef-2018 checkthat! lab on automatic identi cation and veri cation of political claims</article-title>
          .
          <source>In: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          . pp.
          <volume>372</volume>
          {
          <fpage>387</fpage>
          . Springer (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Nikolov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Da San Martino, G.,
          <string-name>
            <surname>Koychev</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Team Alex at CheckThat! 2020:
          <article-title>Identifying check-worthy tweets with transformer models</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Papanastasiou</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Fake news propagation and detection: A sequential model</article-title>
          .
          <source>Management Science</source>
          <volume>66</volume>
          (
          <issue>5</issue>
          ),
          <year>1826</year>
          {
          <year>1846</year>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Passaro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bondielli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcelloni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>UNIPI-NLE at CheckThat! 2020: Approaching fact checking from a sentence similarity perspective through the lens of transformers</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.: Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <issue>1543</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Popat</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mukherjee</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Strotgen, J.,
          <string-name>
            <surname>Weikum</surname>
          </string-name>
          , G.:
          <article-title>Credibility assessment of textual claims on the web</article-title>
          .
          <source>In: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management</source>
          . pp.
          <volume>2173</volume>
          {
          <issue>2178</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Qi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Bolton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Manning</surname>
          </string-name>
          , C.D.:
          <article-title>Stanza: A Python natural language processing toolkit for many human languages</article-title>
          .
          <source>In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations</source>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Reimers</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Sentence-bert: Sentence embeddings using siamese bertnetworks</article-title>
          . arXiv preprint arXiv:
          <year>1908</year>
          .
          <volume>10084</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41.
          <string-name>
            <surname>Reimers</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Making monolingual sentence embeddings multilingual using knowledge distillation</article-title>
          . arXiv preprint arXiv:
          <year>2004</year>
          .
          <volume>09813</volume>
          (
          <issue>04</issue>
          <year>2020</year>
          ), http: //arxiv.org/abs/
          <year>2004</year>
          .09813
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Samadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Talukdar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Veloso</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blum</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Claimeval: Integrated and exible framework for claim evaluation using credibility of sources</article-title>
          .
          <source>In: Thirtieth AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          43.
          <string-name>
            <surname>Shaar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nikolov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Babulkov</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alam</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barron-Ceden</surname>
            ~o,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasanain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suwaileh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haouari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , Da San Martino, G.,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Overview of CheckThat! 2020 English:
          <article-title>Automatic identi cation and veri cation of claims in social media</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          44.
          <string-name>
            <surname>Shearer</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eva Matsa</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <string-name>
            <surname>How Social Media Has Changed How We Consume News</surname>
          </string-name>
          (
          <year>2018</year>
          ), https://www.journalism.org/
          <year>2018</year>
          /09/10/news-use
          <article-title>-acrosssocial-media-platforms-2018/</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          45.
          <string-name>
            <surname>Shu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bernard</surname>
            ,
            <given-names>H.R.</given-names>
          </string-name>
          , Liu, H.:
          <article-title>Studying fake news via network analysis: detection and mitigation</article-title>
          .
          <source>In: Emerging Research Challenges and Opportunities in Computational Social Network Analysis and Mining</source>
          , pp.
          <volume>43</volume>
          {
          <fpage>65</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          46.
          <string-name>
            <surname>Shu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sliva</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Liu, H.:
          <article-title>Fake news detection on social media: A data mining perspective</article-title>
          .
          <source>ACM SIGKDD explorations newsletter 19(1)</source>
          ,
          <volume>22</volume>
          {
          <fpage>36</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          47.
          <string-name>
            <surname>Sicilia</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giudice</surname>
            ,
            <given-names>S.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pei</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pechenizkiy</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Soda</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Twitter rumour detection in the health domain</article-title>
          .
          <source>Expert Systems with Applications</source>
          <volume>110</volume>
          , 33{
          <fpage>40</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          48.
          <string-name>
            <surname>Singhal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>R.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumaraguru</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Satoh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Spotfake: A multi-modal framework for fake news detection</article-title>
          .
          <source>In: 2019 IEEE Fifth International Conference on Multimedia Big Data (BigMM)</source>
          . pp.
          <volume>39</volume>
          {
          <fpage>47</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          49.
          <string-name>
            <surname>Soliman</surname>
            ,
            <given-names>A.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eissa</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>El-Beltagy</surname>
            ,
            <given-names>S.R.</given-names>
          </string-name>
          :
          <article-title>Aravec: A set of arabic word embedding models for use in arabic nlp</article-title>
          .
          <source>Procedia Computer Science</source>
          <volume>117</volume>
          ,
          <issue>256</issue>
          {
          <fpage>265</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          50.
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ounis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Entity detection for check-worthiness prediction: Glasgow terrier at clef checkthat! 2019 (</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          51.
          <string-name>
            <surname>Suykens</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vandewalle</surname>
          </string-name>
          , J.:
          <article-title>Least squares support vector machine classi ers</article-title>
          .
          <source>Neural processing letters 9(3)</source>
          ,
          <volume>293</volume>
          {
          <fpage>300</fpage>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          52.
          <string-name>
            <given-names>Tandoc</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.C.</given-names>
            ,
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.W.</given-names>
            ,
            <surname>Ling</surname>
          </string-name>
          , R.: De ning \
          <article-title>fake news" a typology of scholarly de nitions</article-title>
          .
          <source>Digital journalism 6(2)</source>
          ,
          <volume>137</volume>
          {
          <fpage>153</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          53.
          <string-name>
            <surname>Thuma</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Motlogelwa</surname>
            ,
            <given-names>N.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leburu-Dingalo</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mudongo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : UB ET at CheckThat! 2020:
          <article-title>Exploring ad hoc retrieval approaches in veri ed claims retrieval</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          54.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ma</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jin</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xun</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jha</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
          </string-name>
          , J.: Eann:
          <article-title>Event adversarial neural networks for multi-modal fake news detection</article-title>
          .
          <source>In: Proceedings of the 24th acm sigkdd international conference on knowledge discovery &amp; data mining</source>
          . pp.
          <volume>849</volume>
          {
          <issue>857</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          55.
          <string-name>
            <surname>Wen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>ThunderSVM: A fast SVM library on GPUs and CPUs</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>19</volume>
          ,
          <volume>797</volume>
          {
          <fpage>801</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          56.
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rodrigues</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Novak</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          : Accenture at CheckThat! 2020:
          <article-title>If you say so: Post-hoc fact-checking of claims using transformer-based models</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref57">
        <mixed-citation>
          57.
          <string-name>
            <surname>Wold</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Esbensen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Geladi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Principal component analysis</article-title>
          .
          <source>Chemometrics and intelligent laboratory systems 2(1-3)</source>
          ,
          <volume>37</volume>
          {
          <fpage>52</fpage>
          (
          <year>1987</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref58">
        <mixed-citation>
          58.
          <string-name>
            <surname>Yasser</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kutlu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
          </string-name>
          , T.: bigir at clef 2018:
          <article-title>Detection and veri cation of check-worthy political claims</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref59">
        <mixed-citation>
          59.
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Resnick</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mei</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          :
          <article-title>Enquiring minds: Early detection of rumors in social media from enquiry posts</article-title>
          .
          <source>In: Proceedings of the 24th international conference on world wide web</source>
          . pp.
          <volume>1395</volume>
          {
          <issue>1405</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref60">
        <mixed-citation>
          60.
          <string-name>
            <surname>Zubiaga</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aker</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bontcheva</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liakata</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Procter</surname>
          </string-name>
          , R.:
          <article-title>Detection and resolution of rumours in social media: A survey</article-title>
          .
          <source>ACM Computing Surveys (CSUR) 51(2)</source>
          ,
          <volume>1</volume>
          {
          <fpage>36</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref61">
        <mixed-citation>
          61.
          <string-name>
            <surname>Zubiaga</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liakata</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Procter</surname>
          </string-name>
          , R.:
          <article-title>Exploiting context for rumour detection in social media</article-title>
          .
          <source>In: International Conference on Social Informatics</source>
          . pp.
          <volume>109</volume>
          {
          <fpage>123</fpage>
          . Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref62">
        <mixed-citation>
          62.
          <string-name>
            <surname>Zuo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karakas</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>A hybrid recognition system for check-worthy claims using heuristics and supervised learning</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>