<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>De-Factify</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>S Suryavardan</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shreyash Mishra</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Megha Chakraborty</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Parth Patwa</string-name>
          <email>parthpatwa@g.ucla.edu</email>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anku Rani</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aman Chadha</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aishwarya Reganti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amitava Das</string-name>
          <email>amitava@mailbox.sc.edu</email>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amit Sheth</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manoj Chinnakotla</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Asif Ekbal</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Srijan Kumar</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>IIIT Sri City</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>India</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amazon AI</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Microsoft</string-name>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Carnegie Mellon University</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Georgia Tech</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>IIT Patna</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Stanford</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of California Los Angeles</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of South Carolina</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>Washington</institution>
          ,
          <addr-line>DC</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>2</volume>
      <issue>2</issue>
      <abstract>
        <p>With social media usage growing exponentially in the past few years, fake news has also become extremely prevalent. The detrimental impact of fake news emphasizes the need for research focused on automating the detection of false information and verifying its accuracy. In this work, we present the outcome of the Factify 2 shared task, which provides a multi-modal fact verification and satire news dataset, as part of the DeFactify 2 workshop at AAAI'23. The data calls for a comparison based approach to the task by pairing social media claims with supporting documents, with both text and image, divided into 5 classes based on multi-modal relations. In the second iteration of this task we had over 60 participants and 9 final test-set submissions. The best performances came from the use of DeBERTa for text and Swinv2 and CLIP for image. The highest F1 score averaged for all five classes was</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>81.82%.</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>In recent years, automatic fact-checking has become an important problem in the AI community
due to the increasing prevalence of fraudulent claims masquerading as declarations of reality.
The rapid distribution of news across numerous media sources, particularly on social media</p>
      <p>
        †Work does not relate to position at Amazon.
platforms, has led to the fast development of erroneous and fake content. The fake news
becomes even more harmful during pandemic, elections etc [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. Uncovering misleading
statements before they cause significant harm has become a challenging task. Studies indicate
that a large percentage of the population believes that fake news creates uncertainty, while
only a fraction feels confident in recognizing bogus news. However, the scarcity of available
training data has hindered automated fact-checking eforts.
      </p>
      <p>
        Significant progress has been made with the release of large datasets like [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ], which
include extensive claims and contextual metadata. Although these datasets have contributed to
research advancements, they were purpose-made and may not capture the patterns present in
real-world data efectively. Furthermore, most of the existing datasets focus only on text-based
fake news. To address this limitation, Factify 1 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] introduced a multimodal fact-checking dataset
that consists of original samples with no post-processing or manual data creation involved. The
dataset includes images, textual claims, and reference textual documents/images, facilitating
the exploration of visual cues to enhance fake content detection.
      </p>
      <p>
        Factify 2 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], the latest iteration of the Factify dataset, introduced several enhancements and
new challenges. Factify 2 expands the dataset with an additional 50,000 instances, encompassing
satirical articles that present fake news in a diferent manner. By incorporating satirical content,
it aims to capture the nuanced complexities associated with detecting misinformation in diverse
formats. The Factify datasets addresses the limitations of previous unimodal fact-checking
research eforts by providing a benchmark for researchers to build and evaluate multimodal
fact verification systems. Each data sample in the dataset is labeled with one of five choices:
support, no-evidence, refute (both in text and multimodal components), and satirical. The data is
collected from popular news channels’ Twitter handles in the United States and India, ensuring
its relevance to real-world scenarios. Factify 2 builds upon the success of its predecessor by
expanding the dataset, introducing satirical articles, and creating new challenges for multimodal
fact verification.
      </p>
      <p>This paper presents the findings of the shared task Factify 2, which was organized as part of
the workshop at AAAI 2023. The shared task brought together researchers and practitioners to
evaluate and compare their approaches in multimodal fact verification using the Factify 2 dataset.
Section 2 describes the related work in this domain. We detail the task and the requirements
in 3, followed by the participating teams in section 4. The results of all participating teams’
models are presented in Section 5. Finally, we summarize the task, discuss further research
opportunities, and provide open-ended pointers in Section 7.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Related Work</title>
      <p>
        Multi-modal learning: Researchers have looked into a number of models that incorporate both
textual and visual information in the field of multi-modal learning. This includes approaches
that range from concatenation of individual embeddings to attention fusion layers. ViLBERT
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] is one such approach that processes both visual and textual inputs separately. They interact
through co-attentional transformer layers while extending the BERT architecture. LXMERT
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] builds a large-scale Transformer model with an object relationship encoder, a language
encoder, and a cross-modality encoder, for which the model is trained via pre-training tasks like
masked language modeling, masked object prediction, etc. to learn intra and cross modality
relationships. Learning Joint embedddings using object-word pairs is proposed by [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ].
Other notable work includes OFA [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], BLIP [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], ALIGN [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], CLIP[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], etc.
      </p>
      <p>
        Unimodal fake news detection: Many workshops and shared tasks have been conducted
on unimodal/text based fake news detection and fact verification [
        <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
        ]. Datasets like LIAR
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Covid-19 fake news [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], CREDBANK [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] use claims from sources such as social media and
PolitiFact and are labeled into fake/real or other fine-grained categories. FEVER [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] provides
claims with supporting wikipedia articles that are classified into ”Supported”, ”Refuted” or
”NotEnoughInfo”. Fever and Feverous[
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] use artificial claims that were generated or modified
for the task, which may not be useful in real world scenarios. Recently, bert-based methods
have been very popular in to tackle text based fake news detection [
        <xref ref-type="bibr" rid="ref19 ref23 ref24">19, 23, 24</xref>
        ]. While unimodal
fake news detection might be a good starting point, there is information loss by not utilizing
multiple modalities, which may be critical in fake news detection.
      </p>
      <p>
        Multimodal fake news detection: Although under-explored, some of the recent works aim
for multimodal fake news detection. The Fakeddit dataset [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] has 1 million samples containing
text-image pairs classified into fine-grained labels as well as high-level labels. FakeNewsNet
[
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], collects data from fact checking websites focused on mitigation and spreading for fake
news, and provide social context and dynamic information along with the news. Papadopoulou
et al. [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] provide a dataset of 380 verified and debunked videos. Methods to tackle this problem
have been proposed in [
        <xref ref-type="bibr" rid="ref28 ref29 ref30 ref31">28, 29, 30, 31</xref>
        ]. For a detailed survey on multimodal fake news detection,
please refer [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ].
      </p>
      <p>
        Factify 1: The previous iteration, Factify 1 shared task [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] at AAAI 2022 was conducted
on a multi-modal dataset of having 50k instances. Each data point includes image-text pair
for a claim and supporting document, and each claim-document pair is classified into Support,
insuficient, neutral. Factify 2 releases additional 50k instances which include data points taken
from satirical news.
      </p>
    </sec>
    <sec id="sec-4">
      <title>3. Task Details</title>
      <p>
        The Factify 2 task is designed similar to the previous iteration of the task [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The task aims
to detect fake news with an automated model that is able to classify a given claim based on
whether the document entails it or refutes it. Thus, the dataset has 50,000 data points such that
each sample has a claim-document pair. Here, the claim is defined as a social media post and
the document is an article surrounding the claim. Due to the nature of social media content,
the dataset encourages the use of textual and visual features by providing a text and an image
for every claim and document. The entailment between the four data sources, namely claim
image, claim text, document image and document text, are used to define the categories that
the data is to be classified into. This is also shown in Figure 2.
      </p>
      <p>
        The task description and access to the dataset is available in the Factify 2 task page at
https://codalab.lisn.upsaclay.fr/competitions/8275.
3.1. Data
The dataset [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] consists of claim-document pairs and is curated by combining data from Twitter,
Fact Checking websites and Satirical news websites 1. The claims were extracted from tweets
from Hindustan Times, ANI, ABC and CNN, with their corresponding document text extracted
from the news articles linked to the tweets. Based on metrics like textual and image similarity,
the collected samples were classified into the Support and Neutral categories. Samples for the
refute category were collected from fact checking websites - the fake news was selected as the
claim
      </p>
      <p>
        and the article contents were chosen as the corresponding document. Similarly, Satirical
websites were scraped for text and image, and as the article supports the satirical i.e. false claim,
they were added to the Support category. Satirical headlines were also searched on Google to
ifnd news articles that Refuted or were Insuficient for the claim. The train-validation-test split
of the 50,000 samples is 70:15:15, with each category having the same number of samples. The
train and validation split was provided to the participants and the test set was hidden and used
for evaluation. A more in-depth description of the collection process is presented in the data
paper for the task [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <sec id="sec-4-1">
        <title>3.2. Baseline</title>
        <p>
          The baseline for the Factify 2 dataset is presented in [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. It uses a pre-trained textual and visual
feature extraction models before concatenating the features and passing it to a classifier, as
shown in Figure 3. The data paper compares few pre-trained models for feature extraction and
we finds that the SBERT-mpnet model for text and ViT for image gave the best performance
with a F1 score of 64.99%. Please refer to [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] for more details of the baseline.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>3.3. Evaluation</title>
        <p>The Factify 2 dataset has 5 classes, that the data points are to be categorised into: “Support_Text“,
“Support_Multimodal“, “Insuficient_Text“, “Insuficient_Multimodal“ and “Refute“. To compute
the performance of the classification models, we use F1 score between the ground truth labels
and the predicted labels.</p>
        <p>1 =
2 ×    ×</p>
        <p>+ 
   =
     +    
    
,  =
     +    
    
We calculate the weighted average F1 score for individual classes and also for all samples
together. These scores are computed for all the participating systems and the baseline, as shown
in Table 1. The participants were allowed to make 3 submission and the best score out of
them was used for the leaderboard. The final F1 score across all classes is used to rank the
participants.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Participating systems</title>
      <p>We had over 60 registrations, of which 9 teams provided their final predictions on the test
set and 7 teams made paper submissions to the workshop. We provide an overview of the
approaches used by the participating teams in this section.</p>
      <p>
        Triple-Check [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ] propose a model with pre-trained DeBERTa[
        <xref ref-type="bibr" rid="ref35">35</xref>
        ] for text and Swinv2[
        <xref ref-type="bibr" rid="ref36">36</xref>
        ]
for image embeddings, that are combined using a co-attention fusion block. Their novelty is
in the use of an adapter to train only a few parameters in their model. Features such as text
length, OCR etc. are also used for the final classification.
      </p>
      <p>
        INO [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] use a structure coherence-based approach with components such as textual feature
similarity, textual semantic similarity, text length and image similarity. CLIP [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], S-BERT [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ]
and the ROUGE [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ] are used for text features and ResNet for image features. These components
are used for the final classification through a Random forest classifier.
      </p>
      <p>
        Logically [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ] call their architecture a cross-modal veracity prediction model. They use
both uni-modal embeddings from Word2Vec and multi-modal embeddings from CLIP for their
classification, after passing them through a multi-head attention layer. The document being fed
to this model is from a evidence-retrieval stage, where they rank the paragraphs in document
and use the top-K passages.
      </p>
      <p>
        Zhang [
        <xref ref-type="bibr" rid="ref41">41</xref>
        ] use uni-modal embeddings from pre-trained DeBERTa and ViT, passed to signed
attention layers and a feed-forward network, as well as multi-modal embeddings from CLIP for
their classification. This defined as three modules, namely a text-semantic feature module, a
image-semantic feature module and a text-image correlation module, together referred to as
’UFCC’.
      </p>
      <p>
        gzw [
        <xref ref-type="bibr" rid="ref42">42</xref>
        ] experimented with a inter-modality and intra-modality fusion of textual and visual
embeddings using the co-attention mechanism for their classification model. They refer to
this architecture as Multimodal Attention and Fusion Network (MAFN). They use diferent
pre-trained models within MAFN and make predictions with an ensemble.
      </p>
      <p>
        coco [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ] use an embeddings layer to combine individual embeddings from pre-trained
DeBERTa for text and DeiT[44] for images. This embedding is then passed to bi-directional
hybrid attention mechanism to fuse the claim and document data. The final classification uses
an ensemble of two model pipelines.
      </p>
      <p>Noir [45] obtain image features from pre-trained CLIP-ViT and text features from multilingual
CLIP-ViT. Claim and document embeddings are concatenated before being fed to a multi-headed
attention layer. Then they train a MLP for classification with the output features for image and
text concatenated.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Results</title>
      <p>
        For Factify 2, the participants were allowed to make a maximum of three submissions on the
test set. The Final F1 score was used to decide which of the three prediction files is the best
performing for each participant. Based on that criteria, we present the results of the 9 teams
that made test-phase submissions in Table 1. The table shows the weighted average F1 score
for the 5 individual classes as well as for the overall test set. All teams except one, improved
on the baseline by a minimum of 6.3%. The best performing team is Triple-Check [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ] with a
Final Score of 81.82%, which is about 26% higher than the baseline. No single team performed
better than other team in all categories, which shows that the problem is challenging and needs
further research attention.
      </p>
      <p>Despite adding the satire news articles, the results for the Refute category are very high,
similar to the Factify 1. This may be due to the fact that the document collected from of fact
checking websites are diferent to the writing in formal news articles. While only having a
marginal diference, Support_Text has the lowest scores compared to other categories. This is
further emphasised by the observation that, of samples that were predicted incorrectly by all
participants, Support_Text has the highest occurrence, while Refute followed by Insuficient_Text
occur the least. There are some confusing data points where all the systems failed to predict
the correct class. Some such examples are displayed in Figure 4. From the first example, we can
see that the document is unrelated to claim, but both of them have a similar drone image, hence
leading to confusion. Similarly, in the second example, both the document and claim are about
schools re-opening but in diferent locations, which the models failed to capture.</p>
    </sec>
    <sec id="sec-7">
      <title>6. Conclusion and Future Work</title>
      <p>In this paper, we outline and discuss the Factify 2 shared task on multi-modal fact verification.
The participants utilized various techniques for text embeddings, including DeBERTa, CLIP,
S-BERT, ROUGE and Word2Vec. The image embeddings were extracted through Swinv2, ResNet,
CLIP, ViT and DeiT. Similar to Factify1, ensemble techniques were popular, and some teams
opted to use multiple embeddings to capture features. The shared task described in this work
seeks to identify fake news, but we are far from our goal. No single system could excel in
all categories and there are few examples where all the systems failed, which highlights the
challenges of the task. One possible direction for further research includes using synthetic
fake news data that matches the general data distribution, thus adding complexity to the refute
category.
workshop at AAAI, 2023.
[44] H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, H. Jégou, Training data-eficient
image transformers &amp; distillation through attention, 2021. arXiv:2012.12877.
[45] Z. Zhang, H. Yang, C. Huang, Y. Zhang, Team Noir at Factify 2: Multimodal fake news
detection with pre-trained clip and fusion network, in: De-Factify 2 workshop at AAAI,
2023.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>N.</given-names>
            <surname>Karimi</surname>
          </string-name>
          , J. Gambrell,
          <article-title>hundreds die of poisoning in iran as fake news suggests methanol cure for virus, 2020</article-title>
          . URL: https://www.timesofisrael.
          <article-title>com/ hundreds-die-of-poisoning-in-iran-as-fake-news-suggests-methanol-cure-for-virus/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gandhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kothari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sukumaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chharia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adhikesaven</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rathod</surname>
          </string-name>
          , I. Nandutu,
          <string-name>
            <surname>S. TV</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Misra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Murali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jakimowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Iyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mehra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radunsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Katiyar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>James</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dalal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Advani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dhaliwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Raskar</surname>
          </string-name>
          ,
          <article-title>Challenges of equitable vaccine distribution in the covid-</article-title>
          19 pandemic,
          <year>2022</year>
          . arXiv:
          <year>2012</year>
          .12263.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Morales</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Barbar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gandhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Landage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vats</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kothari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sukumaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mathur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Misra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          , S. T. V.,
          <string-name>
            <given-names>M.</given-names>
            <surname>Arseni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Advani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jakimowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Katiyar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mehra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Iyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Murali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mahindra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dmitrienko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangavarapu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Penrod</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Raskar</surname>
          </string-name>
          , Covid-19 tests gone rogue: Privacy, eficacy, mismanagement and misunderstandings,
          <year>2021</year>
          . arXiv:
          <volume>2101</volume>
          .
          <fpage>01693</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Muhammed</surname>
          </string-name>
          <string-name>
            <surname>T</surname>
          </string-name>
          , S. K. Mathew,
          <article-title>The disaster of misinformation: a review of research in social media</article-title>
          ,
          <source>International Journal of Data Science and Analytics</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>271</fpage>
          -
          <lpage>285</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>W. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          , ”
          <article-title>liar, liar pants on fire”: A new benchmark dataset for fake news detection</article-title>
          ,
          <source>arXiv preprint arXiv:1705.00648</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mitra</surname>
          </string-name>
          , E. Gilbert,
          <article-title>Credbank: A large-scale social media corpus with associated credibility annotations</article-title>
          ,
          <source>in: ICWSM</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>I.</given-names>
            <surname>Augenstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lioma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Chaves</given-names>
            <surname>Lima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Grue</given-names>
            <surname>Simonsen</surname>
          </string-name>
          ,
          <article-title>Multifc: A real-world multi-domain dataset for evidence-based fact checking of claims</article-title>
          , in: EMNLP, Association for Computational Linguistics,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reganti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ekbal</surname>
          </string-name>
          , et al.,
          <article-title>Factify: A multi-modal fact verification dataset</article-title>
          , in: De-Factify workshop at AAAI,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reganti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chadha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chinnakotla</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ekbal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Factify 2: A multimodal fake news and satire news dataset</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Parikh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <article-title>Lxmert: Learning cross-modality encoder representations from transformers</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1908</year>
          .07490.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Gunti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramamoorthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
          </string-name>
          ,
          <article-title>Memotion analysis through the lens of joint embedding (student abstract)</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>36</volume>
          (
          <year>2022</year>
          )
          <fpage>12959</fpage>
          -
          <lpage>12960</lpage>
          . URL: https://ojs.aaai.org/index.php/AAAI/article/view/ 21616. doi:
          <volume>10</volume>
          .1609/aaai.v36i11.
          <fpage>21616</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>V.</given-names>
            <surname>Krishna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramamoorthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chadha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
          </string-name>
          , Imaginator:
          <article-title>Pre-trained image+text joint embeddings using word-level grounding of images</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2305</volume>
          .
          <fpage>10438</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Men</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yang</surname>
          </string-name>
          , Ofa:
          <article-title>Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework</article-title>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2202</volume>
          .
          <fpage>03052</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hoi</surname>
          </string-name>
          ,
          <article-title>Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation</article-title>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2201</volume>
          .
          <fpage>12086</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Parekh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Duerig</surname>
          </string-name>
          ,
          <article-title>Scaling up visual and vision-language representation learning with noisy text supervision</article-title>
          ,
          <year>2021</year>
          . arXiv:
          <volume>2102</volume>
          .
          <fpage>05918</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Krueger</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Learning transferable visual models from natural language supervision</article-title>
          ,
          <year>2021</year>
          . arXiv:
          <volume>2103</volume>
          .
          <fpage>00020</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <article-title>Fever: a large-scale dataset for fact extraction and verification</article-title>
          , arXiv preprint arXiv:
          <year>1803</year>
          .
          <volume>05355</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bhardwaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Guptha</surname>
          </string-name>
          , G. Kumari,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. PYKL</given-names>
            ,
            <surname>A. Das</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ekbal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          , T. Chakraborty,
          <article-title>Overview of constraint 2021 shared tasks: Detecting english covid-19 fake news and hindi hostile posts</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Combating Online Hostile Posts in Regional Languages during Emergency Situation (CONSTRAINT)</source>
          , Springer,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pykl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Guptha</surname>
          </string-name>
          , G. Kumari,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ekbal</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <article-title>Fighting an infodemic: Covid-19 fake news dataset, in: Combating Online Hostile Posts in Regional Languages during Emergency Situation (CONSTRAINT) 2021</article-title>
          , Springer,
          <year>2021</year>
          , p.
          <fpage>21</fpage>
          -
          <lpage>29</lpage>
          . URL: http://dx.doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -73696-
          <issue>5</issue>
          _3. doi:
          <volume>10</volume>
          . 1007/978- 3-
          <fpage>030</fpage>
          - 73696-
          <issue>5</issue>
          _
          <fpage>3</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mitra</surname>
          </string-name>
          , E. Gilbert,
          <article-title>Credbank: A large-scale social media corpus with associated credibility annotations</article-title>
          ,
          <source>in: Proceedings of the international AAAI conference on web and social media</source>
          , volume
          <volume>9</volume>
          ,
          <year>2015</year>
          , pp.
          <fpage>258</fpage>
          -
          <lpage>267</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>R.</given-names>
            <surname>Aly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Schlichtkrull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vlachos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Cocarascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <article-title>The fact extraction and verification over unstructured and structured information (feverous) shared task</article-title>
          ,
          <source>in: Proceedings of the Fourth Workshop on Fact Extraction and VERification (FEVER)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>B.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <article-title>Transformerbased language model fine-tuning methods for covid-</article-title>
          19
          <source>fake news detection</source>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2101</volume>
          .
          <fpage>05509</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Kaliyar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Goswami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Narang</surname>
          </string-name>
          , Fakebert:
          <article-title>Fake news detection in social media with a bert-based deep learning approach</article-title>
          ,
          <source>Multimedia Tools and Applications</source>
          <volume>80</volume>
          (
          <year>2021</year>
          )
          <fpage>11765</fpage>
          -
          <lpage>11788</lpage>
          . URL: https://doi.org/10.1007/s11042-020-10183-2. doi:
          <volume>10</volume>
          .1007/ s11042- 020- 10183- 2.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nakamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Levy</surname>
          </string-name>
          , W. Y. Wang, r/fakeddit:
          <article-title>A new multimodal benchmark dataset for ifne-grained fake news detection</article-title>
          , arXiv preprint arXiv:
          <year>1911</year>
          .
          <volume>03854</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>K.</given-names>
            <surname>Shu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mahudeswaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          , H. Liu,
          <article-title>Fakenewsnet: A data repository with news content, social context and spatialtemporal information for studying fake news on social media</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1809</year>
          .01286.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>O.</given-names>
            <surname>Papadopoulou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Kompatsiaris</surname>
          </string-name>
          ,
          <article-title>A corpus of debunked and verified user-generated videos</article-title>
          ,
          <source>Online information review 43</source>
          (
          <year>2019</year>
          )
          <fpage>72</fpage>
          -
          <lpage>88</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>Multimodal fusion with co-attention networks for fake news detection, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>2560</fpage>
          -
          <lpage>2569</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .findings-acl.
          <volume>226</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          . findings- acl.226.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <surname>H. Zhang,</surname>
          </string-name>
          <article-title>Multimodal fake news detection via progressive fusion networks</article-title>
          ,
          <source>Information processing &amp; management 60</source>
          (
          <year>2023</year>
          )
          <fpage>103120</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <article-title>Multimodal fake news detection through data augmentation-based contrastive learning</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>136</volume>
          (
          <year>2023</year>
          )
          <fpage>110125</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yadav</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gaba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Budhiraja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <surname>Etma:</surname>
          </string-name>
          <article-title>Eficient transformerbased multilevel attention framework for multimodal fake news detection</article-title>
          ,
          <source>IEEE Transactions on Computational Social Systems</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cresci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D. S.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Firooz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>A survey on multimodal disinformation detection</article-title>
          ,
          <source>in: Proceedings of the 29th International Conference on Computational Linguistics</source>
          ,
          <source>International Committee on Computational Linguistics</source>
          , Gyeongju, Republic of Korea,
          <year>2022</year>
          , pp.
          <fpage>6625</fpage>
          -
          <lpage>6643</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .coling-
          <volume>1</volume>
          .
          <fpage>576</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>P.</given-names>
            <surname>Patwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryavardan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reganti</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sheth</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ekbal</surname>
          </string-name>
          , et al.,
          <article-title>Benchmarking multi-modal entailment for fact verification</article-title>
          , in: Proceedings of De-Factify: Workshop on Multimodal Fact Checking and
          <article-title>Hate Speech Detection</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>W.-W.</given-names>
            <surname>Du</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-W. Wu</surname>
          </string-name>
          , W.-Y. Wang, W.-C.
          <article-title>Peng, Team Triple-check at factify 2: Parametereficient large foundation models with feature representations for multi-modal fact verification</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , W. Chen, Deberta:
          <article-title>Decoding-enhanced bert with disentangled attention</article-title>
          ,
          <year>2021</year>
          . arXiv:
          <year>2006</year>
          .03654.
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Guo</surname>
          </string-name>
          , Swin transformer v2:
          <article-title>Scaling up capacity</article-title>
          and resolution,
          <year>2022</year>
          . arXiv:
          <volume>2111</volume>
          .
          <fpage>09883</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          , INO at Factify 2:
          <article-title>Structure coherence based multimodal fact verification</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <surname>C.-Y. Lin</surname>
            ,
            <given-names>ROUGE:</given-names>
          </string-name>
          <article-title>A package for automatic evaluation of summaries, in: Text Summarization Branches Out, Association for Computational Linguistics</article-title>
          , Barcelona, Spain,
          <year>2004</year>
          , pp.
          <fpage>74</fpage>
          -
          <lpage>81</lpage>
          . URL: https://aclanthology.org/W04-1013.
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>P.</given-names>
            <surname>Jordi Verschuuren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Van</given-names>
            <surname>Eeden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Oikonomou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bandhakavi</surname>
          </string-name>
          ,
          <article-title>Logically at Factify 2: A multi-modal fact checking system based on evidence retrieval techniques and transformer encoder architecture</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Liu, E. Sun, Team Zhang at Factify 2:
          <string-name>
            <surname>Unimodal</surname>
          </string-name>
          feature
          <article-title>-enhanced and cross-modal correlation learning for multi-modal fact verification</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Team gzw at Factify 2023:
          <article-title>Multimodal attention and fusion networks for multi-modal fact verification</article-title>
          , in: De-Factify 2 workshop
          <string-name>
            <surname>at</surname>
            <given-names>AAAI</given-names>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>K.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Luo</surname>
          </string-name>
          , R. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>Team coco at Factify 2: Optimizing modality representation and multimodal fusion for multimodal fact verification</article-title>
          , in: De-Factify 2
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>