<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>UofA-Truth at Factify 2022 : A Simple Approach to Multi-Modal Fact-Checking</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Abhishek Dhankar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Osmar R. Zaïane</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francois Bolduc</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Alberta</institution>
          ,
          <addr-line>Edmonton, Alberta, T6G 2E8</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Identifying fake news is a very dificult task, especially when considering the multiple modes of conveying information through text, image, video and/or audio. We attempted to tackle the problem of automated misinformation/disinformation detection in multi-modal news sources (including text and images) through our simple, yet efective, approach in the FACTIFY shared task at De-Factify@AAAI2022. Our model produced an F1-weighted score of 74.807%, which was the fourth best out of all the submissions. In this paper we will explain our approach to undertake the shared task.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;fake news</kwd>
        <kwd>multi-modal</kwd>
        <kwd>De-Factify@AAAI2022</kwd>
        <kwd>FACTIFY shared task</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Humankind has dealt with misinformation since time immemorial [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, never in
human history have people had access to the amount of information that they have today. The
Internet is the primary reason for easy access to this information. It has given people the ability
to access information from all over the world and from innumerable sources. However, this
deluge of information has brought with it the problem of misinformation/disinformation/fake
news. Never before have we had more eficacious means to disseminate deceptive fallacies and
falsehoods that are unfortunately believed and are wrongfully, and sometimes dangerously
impacting people.
      </p>
      <p>
        While there are many definitions of Fake News, for this paper Fake News can be defined as a
news piece, social media post, etc., which contains claim(s) that can be refuted by information
put out by “reputable organizations”. Such organizations may include, but are not limited to,
government bodies, news outlets that score high on Media Bias/Fact Check’s Factual Reporting
scale [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] or professional fact-checking organizations which are verified signatories of the
International Fact-Checking Network (IFCN) code of conduct [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This definition of fake news
of-loads the responsibility of determining what exactly fake news is, on to expert fact-checkers
or domain experts, and allows Artificial Intelligence (AI) to deal with the more manageable
problem of determining whether claim(s) made in a news piece is entailed, not entailed or
refuted by a corresponding news piece from a reputable source.
      </p>
      <p>
        Fake news can cause real world harm as is being seen during the COVID-19 pandemic:
misinformation has led to vaccine hesitancy, which is directly tied to increased chances of
mortality due to COVID-19 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Fact-checking or determining whether a news piece contains
fake claims is the first step in countering such fake news. Furthermore, it is not only important
to detect and counteract fake news, but to do so in a timely manner. Given the large amount of
information generated on social media sites every-day and the time constraints that online
factchecking operates under, it is imperative that automated methods of misinformation detection
are developed to aid in the manual fact-checking of fake news.
      </p>
      <p>In general, the information generated and distributed on the Internet is multi-modal, i.e.,
consisting of text, images, audio-visual, etc. Often times information is conveyed via a combination
of two or more modes, for instance, memes, pieces of information rapidly spread among users,
are often a combination of text and image/short video, where text is overlaid on the image or
short video (also called a gif). Thus, an automated method should be able to take advantage of
all the modes of information available to fact-check a claim.</p>
      <p>
        The shared task FACTIFY, in conjunction with the AAAI conference, attempts to aid in the
development of automated multi-modal fact-checking by introducing a dataset that consists of
multi-modal claims and corresponding supplementary information or documents, using which
said claims need to be fact-checked [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Each multi-modal “claim” consists of a short sentence
or phrase and an associated image, which may or may not have overlaid text. The corresponding
supplementary information or “document” consists of a sentence or sentences from a reputable
source with an accompanying image. The task is to determine whether the claim text and claim
image are individually entailed, not entailed, or refuted by the corresponding document text
and image pair. Depending on that, and defined by the shared task organizers, there are five
possible labels that each claim text and image pair can have:
• Support_Multimodal: Both claim text and image are entailed
• Support_text: Text is entailed, but the image is not
• Insuficient_Multimodal: Claim text is not entailed, but claim image is
• Insuficient_Text: Neither claim text nor claim image is entailed
• Refute: Both claim text and image are refuted
      </p>
      <p>Our team “UofA-Truth” participated in the shared task and secured the 4ℎ position with a
weighted F1-score of 74.807%, just ≈ 2 F1 points behind the top submission. In this paper, we
shall describe our simple yet efective automated fact-checking model.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        The dataset used to train and test our model was released under the shared task FACTIFY, which
is a part of the workshop De-Factify at the AAAI 2022 conference [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The dataset consists of a
total of 50, 000 claim and document pairs, which are divided into train, validation and test sets
of sizes 35, 000(70%), 7, 500(15%) and 7, 500(15%) respectively.
      </p>
      <p>
        The entailment aspect of the shared task is similar to “Stance Detection”, which can be defined
as the classification of the stance of the producer of a news piece concerning an unverified
claim [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In the context of the shared task, the unverified claim is the claim text and image
pair, and the news piece is the document text and image pair.
      </p>
      <p>
        Stance Detection is an important part of Fake News detection and was notably used in the
Fake News Challenge - 1 (FNC-1) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This challenge was similar to the FACTIFY shared task,
except FNC-1 only dealt with text entailment or stance detection, unlike FACTIFY which deals
with multi-modal entailment. FNC-1 introduced a dataset that consisted of a headline and a
body of text, which may be from the same article or diferent articles. Depending on the stance
of the body of text concerning the headline, the text-headline pairs were to be classified into
any of the following classes:
• Agrees: The body of text agrees with the claim(s) made in the headline
• Disagrees: The body of text disagrees with the claim(s) made in the headline
• Discusses: The body of text and headline are referring to the same subject, but the body
does not take any stance or position on the claim(s) made in the headline
• Unrelated: The body of text is not related to the claim(s) being made in the headline
FACTIFY’s not-entail class can be considered similar to a combination of Unrelated and
Discusses classes of FNC-1, while entails and refutes classes can be considered similar to
FNC-1’s Agrees and Disagrees classes respectively.
      </p>
      <p>
        This similarity between the two tasks led us to draw inspiration from the UCL Machine
Reading team’s submission to the FNC-1’s challenge, which performed 3 best among the 50
submissions to the challenge [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In their submission the UCL team, Riedel et al., describe their
approach as a “simple but tough-to-beat baseline” for stance detection. As explained above,
there are two inputs for this task - a headline and a body of text. Riedel et al. calculated the
Term Frequency (TF) vectors and Term Frequency - Inverse Document Frequency (TF-IDF) for
both the headline and the body of text-based on the 5, 000 most frequent words. The TF vectors
of the two inputs are concatenated with the result of the cosine similarity between the TF-IDF
vectors of the headline and body, as shown in Figure 1. The resultant vector of length 10, 001 is
then fed as input into a shallow Multi Layer Perceptron (MLP) network, which has a softmax
output of length four, one for each class in the FNC-1 task.
      </p>
      <p>
        Given the similarity of the tasks being solved in FNC-1 and FACTIFY, we adopted the manner
of the concatenation of the cosine similarity and vector representations of the header and
body as explained in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Instead of using TF for vector representations and TF-IDF for cosine
similarity calculation, we used Sentence-BERT in lieu of both TF and TF-IDF vectors to determine
entailment between the claim text and document text [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Sentence BERT is a transformer
based model which has been specifically trained to represent sentences and paragraphs and
therefore may be better for sentence representation than other methods which usually compute
the average across vector representations of individual words to obtain sentence or paragraph
representations. To determine entailment between claim and document images, we used a
pre-trained instantiation of the Xception architecture [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] available in Keras [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], which had
been trained on JFT-300M dataset [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        The FNC-1 challenge had two other submissions which performed better than [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], however,
we concluded that those were more complicated architectures and might hamper the scalability
and time complexity of our model. For instance, Pan et al., who submitted the winning
model [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], had an ensemble model which consisted of a deep learning model and a tree based
ensemble model as implemented in Xgboost [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. The outputs of the two models were weighted
equally to produce the final predictions. Hanselowski et al. also implemented an ensemble model
which consisted of five Neural Network models. The final prediction was made through majority
voting. Despite the increased complexity, Team UCL Machine Reading’s model performance
was within ≈ 1 point of the top two submissions.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The FACTIFY shared task’s classes (Support_Multimodal, Support_text, Insuficient_Multimodal,
Insuficient_Text, Refute) are composed of a combination of text and image entailment classes.
For instance, if text entailment, non-entailment, and refutation are represented by  _0,  _1,
 _2, and image entailment, non-entailment and refutation are represented by ℐ_0, ℐ_1, ℐ_2
respectively; then the shared task’s classes can be reformulated as a combination of text and
image entailment labels as shown in Table 1. It is important to note here that all combinations
of text entailment labels and image entailment labels are not present in Table 1. For instance,
combinations such as  _0 &amp; ℐ_2 do not exist. The lacking combinations are treated when
consolidating the labels after classification. This is explained in Section 3.4.</p>
      <p>It can be clearly seen that the shared task can now be broken down into two sub-tasks;
namely, text entailment and image entailment, where text entailment consists of classes  _0,
 _1 and  _2, and image entailment consists of classes ℐ_0, ℐ_1, ℐ_2. These new classes
are the combination of the original class labels as shown in Table 2 and Table 3 for the text
entailment and image entailment tasks respectively. Once the dataset is rearranged according
to the sub-task labels, we end up with one dataset for each sub-task.</p>
      <p>We now define Text Entailment as a task of predicting the document text’s stance towards the
claim text, and Image Entailment as a task of predicting the document image’s stance towards
the claim image.</p>
      <sec id="sec-3-1">
        <title>3.1. Preprocessing</title>
        <p>
          Image preprocessing is done by resizing all the images to (256, 256, 3) size with bilinear
interpolation as implemented in image_dataset_from_directory in Keras [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Thereafter, all the pixel
values are scaled to a range of 0 to 1.
        </p>
        <p>
          Text preprocessing involves removing URLs from all claim and document texts with the help
of the Preprocessor library [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Vector Representations</title>
        <p>The preprocessed inputs (text and images) need to be converted into vector representations so
that they can be presented as input for a classifier.</p>
        <p>
          The preprocessed images are converted into vectors of size 2048 each, by using the pre-trained
Xception model in Keras [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. This can be achieved by setting include_top attribute to False
and pooling attribute to ‘avg’. Setting include_top to False removes the fully connected layer at
the end of the model and exposes the output of the second to the last layer. Setting pooling to
‘avg’ ensures that a global pooling average is applied to the 3D output of the second last layer
of Xception, to convert it into a 1D output. Since the Xception model has been trained on a
massive dataset for a general image classification task, it can be reasonably assumed that the
output of the second to the last layer captures information that may be useful for downstream
tasks such as image entailment.
        </p>
        <p>
          The preprocessed texts are converted into vectors of length 384 each by using the pre-trained
Sentence-BERT model [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
        <p>The cosine similarity of the vector representations of claim and corresponding document
images is calculated and concatenated in the manner shown in Figure 2. This creates a
concatenated representation for each claim and corresponding document image of size 4097. Similarly,
the concatenated representation of claim and corresponding document text of size 769 is created
through the same procedure of cosine similarity calculation and subsequent concatenation as
shown in Figure 3.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Classifiers</title>
        <p>The vector representations are now used for training the classifiers for the image and text
entailment tasks. Diferent classifiers are used for the image and text entailment tasks.</p>
        <p>As shown in Figure 2 the image entailment classifier consists of a single fully connected
hidden MLP layer of 5000 units, ReLU activation with a dropout probability of 0.5. The output
of this layer then feeds into a fully connected output layer of 3 units, one for each class label
(entailment, non-entailment and refute), and a sigmoid activation function.</p>
        <p>On the other hand, as shown in Figure 3 the text entailment classifier consists of two fully
connected layers of 450 units each, ReLU activation functions, 2 activity regularizers, and a
dropout probability of 0.55 for the first layer and 0.4 for the second layer. The output of the two
hidden layers then feeds into the fully connected output layer 3 units with sigmoid activation.</p>
        <p>The Cross-Entropy loss is calculated after performing the softmax operation on the outputs
of both the classifiers.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Label Consolidation</title>
        <p>The output of the image classifier classifies every pair of claims and document image into one
of the three labels ℐ_0, ℐ_1 and ℐ_2. Similarly, for claim and document text pairs.</p>
        <p>The pairs of image and text entailment labels belonging to the same data-point are
combined and then converted into the original FACTIFY task labels (namely, Support_Multimodal,
Support_text, Insuficient_Multimodal, Insuficient_Text, Refute) according to Table 1.</p>
        <p>However, it is possible that the combination procedure may produce pairs of entailment
labels that do not have any corresponding FACTIFY task label. For instance, ( _0, ℐ_2), ( _1,
ℐ_2), ( _2, ℐ_0), ( _2, ℐ_1), are four such invalid pairs of labels. We thus have to change
such label pairs into valid label pairs. We do so by using a heuristic as described in Table 4(A).
If one of either claim text or claim image is entailed, i.e.,  _0 or ℐ_0, it is unlikely that the
other claim mode will be refuted by the document, hence, the latter’s label needs to be changed
to not-entailed, i.e.,  _1 or ℐ_1. If however, one of the claim text or image is refuted by the
corresponding document then it is unlikely that the other claim mode will have uncertain
entailment, hence the latter’s label should be converted to refuted as well, i.e.,  _2 or ℐ_2.
Thereafter, we can calculate the final weighted F1 accuracy on the Test set.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results &amp; Discussion</title>
      <p>Invalid Label Pair</p>
      <p>Valid Label Pair
( _0, ℐ_2) ( _0, ℐ_0)
( _1, ℐ_2) ( _1, ℐ_1)
( _2, ℐ_0) ( _2, ℐ_2)
( _2, ℐ_1) ( _2, ℐ_2)
B: New Invalid to Valid Label Pair Conversion
Our team, UofA-Truth, secured the 4ℎ position on the leaderboard, with an F1-score of 74.807%
on the final evaluation. However, the confusion matrix, shown in Figure 4, reveals more
ifne-grained details about our model’s performance on the test set.</p>
      <p>The model performed worst on the Insuficient_Multimodal category. It can be clearly
seen from the matrix that a large number (304) of data points with ground truth
Insuficient_Multimodal were incorrectly classified as Support_Multimodal. Since the only diference
between the two classes is text entailment, it is possible that the model maybe unable to
differentiate between text entailment and non-entailment. This may be because the claim and
document texts might have common words or might even talk about tangential or similar topics,
but do not reach the threshold of text entailment.</p>
      <p>Furthermore, the model did not perform well on the Support_text class. A significant number
(202) of data-points belonging to Support_Text were misclassified as Support_Multimodal.
Again, given that the only diference between the two classes is in image entailment, it follows
that a model would find it hard to diferentiate between the two. Similarly, for the Support_Text
and Insuficient_Text.</p>
      <p>The model performs best on the Refute class despite the fact that the said class had the fewest
data points in the training set. Very few of its data points are misclassified as other classes, and
vice-versa. This may be because a significant number of the data points belonging to the Refute
class have been taken from fact-checking websites. This fact may set such data points apart
from other claim-document pairs. For instance, document images and corresponding claim
images of the Refute class tend to be identical because fact-checking websites almost always
provide a screenshot of the fake news/social media posts they debunk in their articles. They
may even overlay images of news pieces they fact-check with a digital stamp, indicating their
logo or whether the news piece was true or fake. They usually clearly state the gist of the fake
news they debunk, at the beginning of every article, often quoting said fake news verbatim.
Such peculiarities may make data points belonging to the Refute class easy to discern.</p>
      <p>Heuristics mentioned in Section 3.4 can be changed to improve the weighted F1-score on
the test set. It is possible that the image entailment model merely learns to determine the
similarity between claim and document image pairs. Thus, it may be better to have a heuristic
that changes the invalid label pairs into valid label pairs by changing the image entailment label
to be the same as the text entailment label. Therefore, after the competition results, we changed
the heuristic for invalid label pair to valid label pair conversion as per the new heuristics shown
in Table 4(B). These modified heuristics improve the final F1-score from 74.807% to 75.183%.
As can be seen by the confusion matrix in Figure 5, the new heuristic reduces the classification
accuracies of the Insuficient_Multimodal and Support_Text classes, for the benefit of the other
classes. Other than that, the overall dynamics remain the same as in Figure 4. It could be possible
to continue adjusting these heuristics to obtain even better results but have not experimented
further.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper, we introduced a simple, yet efective method of multi-modal fake news detection.
We divided the main task into two sub-tasks; namely, text entailment and image entailment.
Thereafter, we used pre-trained Xception network and Sentence-BERT to get vector
representations of images and text respectively. We then used these vector representations for
classifications tasks of image and text entailment by adapting the approach introduced by
Riedel et al. in their submission to the FNC-1 task. Finally, we consolidated the prediction of
the two sub-tasks of image and text entailment to get the final predictions. We used the model
thus created to make predictions on the test set, and our team’s submission achieved the 4ℎ
position on the leader board with a 74.807% weighted F1-score.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was partially funded by a Collaborative Health Research Project grant from the
Canadian Institutes of Health Research (CIHR) and the Natural Sciences and Engineering
Research Council of Canada (NSERC). Osmar Zaiane, a Canada CIFAR AI Chair, is also funded
by the Canadian Institute for Advanced Research (CIFAR).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mansky</surname>
          </string-name>
          ,
          <article-title>The age-old problem of “fake news</article-title>
          ”,
          <year>2018</year>
          . URL: https://www.smithsonianmag. com/history/age
          <article-title>-old-problem-</article-title>
          <string-name>
            <surname>fake-</surname>
          </string-name>
          news-
          <volume>180968945</volume>
          /, accessed on 2021-
          <volume>11</volume>
          -24.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Methodology</surname>
          </string-name>
          ,
          <year>2021</year>
          . URL: https://mediabiasfactcheck.com/methodology/, accessed on 2021-
          <volume>11</volume>
          -24.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>International</surname>
          </string-name>
          fact-checking network,
          <year>2021</year>
          . URL: https://www.poynter.org/ifcn/, accessed on 2021-
          <volume>11</volume>
          -24.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>[4] Verified signatories of the ifcn code of principles, 2021</article-title>
          . URL: https://ifcncodeofprinciples. poynter.org/signatories, accessed on 2021-
          <volume>11</volume>
          -24.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. S.</given-names>
            <surname>Sy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Glenn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Ryan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Morrissette</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Shay</surname>
          </string-name>
          , G. VazquezBenitez,
          <string-name>
            <surname>J. M. Glanz</surname>
            ,
            <given-names>N. P.</given-names>
          </string-name>
          <string-name>
            <surname>Klein</surname>
          </string-name>
          , et al.,
          <article-title>Covid-19 vaccination and non-covid-19 mortality risk-seven integrated health care organizations, united states</article-title>
          ,
          <source>december 14, 2020-july 31</source>
          ,
          <year>2021</year>
          ,
          <source>Morbidity and Mortality Weekly Report</source>
          <volume>70</volume>
          (
          <year>2021</year>
          )
          <fpage>1520</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reganti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ekbal</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Ahuja</surname>
          </string-name>
          ,
          <article-title>Factify: A multi-modal fact verification dataset</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Multimodal Fact-Checking and Hate</source>
          Speech
          <string-name>
            <surname>Detection (DE-FACTIFY)</surname>
          </string-name>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reganti</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ekbal</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Ahuja</surname>
          </string-name>
          ,
          <article-title>Benchmarking multi-modal entailment for fact verification</article-title>
          , in: Proceedings of De-Factify: Workshop on Multimodal Fact Checking and
          <article-title>Hate Speech Detection</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Küçük</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Can</surname>
          </string-name>
          ,
          <article-title>Stance detection: A survey</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>53</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanselowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>PVS</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schiller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Caspelherr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chaudhuri</surname>
          </string-name>
          , C. M. Meyer, I. Gurevych,
          <article-title>A retrospective analysis of the fake news challenge stance-detection task</article-title>
          ,
          <source>in: Proceedings of the 27th International Conference on Computational Linguistics</source>
          , Association for Computational Linguistics, Santa Fe, New Mexico, USA,
          <year>2018</year>
          , pp.
          <fpage>1859</fpage>
          -
          <lpage>1874</lpage>
          . URL: https://aclanthology.org/C18-1158.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , I. Augenstein,
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Spithourakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <article-title>A simple but tough-to-beat baseline for the fake news challenge stance detection task</article-title>
          ,
          <source>arXiv preprint arXiv:1707.03264</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: https://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          , Xception:
          <article-title>Deep learning with depthwise separable convolutions</article-title>
          ,
          <source>in: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1251</fpage>
          -
          <lpage>1258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          , et al.,
          <string-name>
            <surname>Keras</surname>
          </string-name>
          , https://keras.io,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>G.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Distilling the knowledge in a neural network</article-title>
          ,
          <source>arXiv preprint arXiv:1503.02531</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sibley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Baird</surname>
          </string-name>
          ,
          <article-title>Fake News Challenge - Team SOLAT IN THE SWEN</article-title>
          , https: //github.com/Cisco-Talos/fnc-1,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          , C. Guestrin,
          <article-title>XGBoost: A scalable tree boosting system</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '16</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA,
          <year>2016</year>
          , pp.
          <fpage>785</fpage>
          -
          <lpage>794</lpage>
          . URL: http://doi.acm.
          <source>org/10</source>
          .1145/ 2939672.2939785. doi:
          <volume>10</volume>
          .1145/2939672.2939785.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanselowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>PVS</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schiller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Caspelherr</surname>
          </string-name>
          , athene_system, https://github.com/ hanselowski/athene_system,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Özcan</surname>
          </string-name>
          , santiagonasar, Rusty, Rushat,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lopez</surname>
          </string-name>
          , Piotti, Preprocessor, https://github.com/ s/preprocessor,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>