=Paper=
{{Paper
|id=Vol-3199/paper4
|storemode=property
|title=Tyche at Factify 2022: Fusion Networks for Multi-Modal Fact-Checking (short paper)
|pdfUrl=https://ceur-ws.org/Vol-3199/paper4.pdf
|volume=Vol-3199
|authors=Ankesh Raj,Nainesh Hulke,Ali Asgar Saifee,Bharath Raj Siv
|dblpUrl=https://dblp.org/rec/conf/aaai/RajHSS22
}}
==Tyche at Factify 2022: Fusion Networks for Multi-Modal Fact-Checking (short paper)==
12 Tyche at Factify 2022: Fusion Networks for Multi-Modal Fact-Checking Nainesh Hulke1 , Bharath Raj Siva1 , Ankesh Raj1 and Ali Asgar Saifee1 1 Indian Institute of Technology (Indian School of Mines), Dhanbad, India Abstract This paper describes our approach for the multimodal fact-checking (Factify) challenge at AAAI 2022. Fake news has the potential to harm people and society. To combat fake news, fact-checking becomes crucial. False claims are prevalent in both visual and textual forms. Multimodal techniques can merge information from both these modalities. In our approach, we treated this challenge as a multi-class classification task. We extracted textual and visual features from the texts and images, respectively. We used a single BERT module as a text feature extractor for both claim and reference text so that the extracted feature representations show the perspective of the same model on both claim and reference texts rather than two feature representations from two different models. The same goes for the image feature extractor. EfficientNet-B3 was used for extracting image features. Then the features were passed through a proposed fusion module and the classifier. The purpose of the fusion module is to enable the model to learn semantic similarities between texts and images. We achieved the best F1 score of 0.692 on the test set of the FACTIFY dataset. Keywords Multimodal Fact-Checking, Fusion Network, Fake News Detection 1. Introduction News and information spread quickly through the internet and social media. This has also led to a vast increase in the spread of misinformation and fake news. As per a study, Facebook engagements with fake news sites average roughly 70 million per month[1]. Fake news has the potential to harm people and society. This spread has been blamed for incidents ranging from ethnic violence, inter-racial violence, and religious conflicts to mass riots. For example, some studies argue that fake news is gradually becoming an essential aspect of South Africa’s xenophobia discourse[2]. There has been a lot of recent work on fact-checking claims. Early works on fact-checking focused on using textual information extracted from the text of the article, such as statistical text features[3], emotional information[4][5][6], or integrating metadata with the text[7]. Although 1 All authors contributed equally. 2 Team Name: Tyche De-Factify: Workshop on Multimodal Fact Checking and Hate Speech Detection, co-located with AAAI 2022. 2022 Vancouver, Canada Envelope-Open nainesh.18je0513@cse.iitism.ac.in (N. Hulke); bharathraj.18je0220@cse.iitism.ac.in (B. R. Siva); ankesh.18je0122@cse.iitism.ac.in (A. Raj); munni.18je0077@cse.iitism.ac.in (A. A. Saifee) © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) the textual content can be a significant indicator for fake news detection, it is not sufficient when used alone. Some researchers have proposed systems that use the credibility of the pages that post the news[8] or profile characteristics of the users that shared the post to detect the articles that contain manipulated content[9][10]. Online articles and posts usually contain more information in the form of images and social context that can be useful for fake news detection. Online news contains images that generally attract the attention of users. Images in fake and real news may follow different patterns or be modified to attract users’ attention and make them share them. There are several reasons why an image may be deemed fake. In most cases, this involves digital manipulation, e.g., cropping, splicing, etc. However, there are cases when an image is entirely legitimate, but it is published alongside some text that does not reflect its content accurately. Hence, it is essential that a system also exploits information extracted from the images for effective fake news detection. Visual information can complement the textual one for fake news detection. Multimodal systems can merge information from both these modalities. It is also capable of disambiguating the wrong classifications and improving the results using combined modalities. Some researchers have proposed multimodal systems that combine textual and visual information for determining whether an article or a post is fake or not[11][12] or combined textual, visual, and semantic information for fact-checking claims using a multimodal architecture[13]. 2. Related Work Early attempts to use deep learning methods to counter fake news employed data of a single modality. Vo et al. proposed a novel framework, Fact-checking Response Generator (FCRG)[14], to generate a fact-checking response tweet to combat fake news on Twitter. They showed distinguishing linguistic features of fact-checking tweets compared with Normal and Random Replies. Wang et al.[7] introduced a new dataset LIAR for fake news detection and proposed a CNN model to integrate metadata with text. The proposed method outperforms text-only models strictly, suggesting the use of multiple modalities to express the news article’s intent more clearly. News articles often contain images and text; hence, multimodal deep learning methods have produced good results. Wang et al. [15] proposed the EANN (event adversarial neural network) model with three major components: the multimodal feature extractor, the fake news detector, and the event discriminator. The textual and visual latent feature representations are learned and concatenated to form the final multimodal feature representation. Multimodal approaches that use features from the different modalities and feed them to different types of networks have also been successful. Gallo et al.[16] introduced multimodal fusion networks where the text and image feature representations are concatenated and passed through a neural network. The proposed method achieves the state-of-the-art results on the UPMC Food-101 dataset. Recent research in multimodal fake news detection by Giachanou et al.[13] proposed a multi- modal multi-image network for fake news detection. The proposed system combines textual, Claim Text Features Text Feature Extractor Document Text Features Claim Text Fusion Module Document Text Dense Layer Softmax Claim Image Features Image Feature Extractor Document Image Features Claim Image Document Image Figure 1: Model Architecture visual, and semantic information. For the textual representation, BERT-Base [17] was used to capture the underlying semantic and contextual meaning. Image tags were extracted from multiple images that the articles contained using the VGG-16 model [18] for the visual repre- sentation. The semantic information was represented by the image-text similarity calculated using the title and image tags embeddings cosine similarity. All these features are concatenated to make the final prediction. 3. Proposed Approach Our work is based on the FACTIFY dataset [19] [20]. In our approach, we extract the textual and visual features from the claim and reference texts and images, respectively. We propose a fusion module for the model to learn semantic similarities between the texts and images. The extracted features from this module are concatenated and passed through a classifier. Our approach aims to train a model that can use the relationship between the different modalities, i.e., text and image. Our model mainly consists of three different components Figure-1. • Text Feature Extractor • Image Feature Extractor • Fusion Module We used a single text feature extractor for both claim and reference text so that the extracted feature representations show the perspective of the same model on both the texts rather than two feature representations from two different models. A similar argument can be applied to images as well. Claim Text Features Document Text Dense Layer Features Dense Layer Claim Image Features Image Features Image Features Document Document Figure 2: Text_Image_X2 Fusion Module Figure 3: Text_X2_Image_X2 Fusion Module Document Text Dense Layer Features 3.1. Text Feature Extractor We use the pre-trained Bidirectional Encoder Representations from Transformers (BERT) [17] base model because of its ability to tackle a broad set of NLP tasks. We fine-tuned the model for Dense Layer Claim Image Features our specific task of text feature extraction. Specific details about the BERT model architecture, which we used, are given in Appendix A. Claim Text Features 3.2. Image Feature Extractor We wanted to have a model with high accuracy on ImageNet [21] that can be trainable with limited hardware resources, so we used the EfficientNetB3 architecture, pre-trained on ImageNet, and replaced the last dense layer with a dense layer of size 256 followed by a batch normalization layer. The EfficientNetB3 has 12 million parameters and achieves a top-1 accuracy of 81.6% on ImageNet. It balances the trade-off between classification accuracy and computational efficiency for our task. 3.3. Fusion Module The extracted feature outputs are combined using a fusion module. We propose two methods of fusion with different intuitions on the output. • Text_Image_X2 : In this method Figure-2., the claim text features and claim image features are concatenated and passed through a dense layer of size 256 to give the claim features. Similarly, we also get the reference features. This captures the similarity between text and image features of both claim and reference. • Text_X2_Image_X2 : In this method Figure-3, the claim text features and reference text features are concatenated and passed through a dense layer of size 256 to give the text features. Similarly, we also get the image features. This captures the similarity between text features of claim and reference and image features of claim and reference. The output features are then concatenated and passed through a dense layer of size 5. Finally, a Softmax layer is added to get the class probabilities. 4. Experiments 4.1. Data Preprocessing The FACTIFY dataset has both text and images; therefore, preprocessing was performed sepa- rately. 4.1.1. Text Preprocessing First, the text was converted to lowercase, removing all redundant elements, such as ASCII characters and weblinks. The stop words (i.e., articles, connectors, prepositions, and others) were removed since they do not contain any information and the semantics of sentences remain intact. We used the nltk [22] stopwords list to remove stopwords. The words of a sentence were then lemmatized. Lemmatization involves using a vocabulary and morphological analysis of words to remove inflectional endings only and return the base or dictionary form of the word. Thus, the size of the text was reduced, and the semantics of the sentence were intact. Only important words remained. 4.1.2. Image Preprocessing The training data have a variable aspect ratio and size, so we adopted a stage-dependent re- scaling policy. We re-scale an image in the training stage so that the min side is 256 pixels wide while keeping the initial image’s aspect ratio. As a result, the spatial information was not lost, and at the same time, the input to models was within trainable limits. Various other augments from Albumentations [23] were adopted to prevent overfitting. Since all the images are similar to real-world images, all arguments were considered to mirror real-world images. 4.2. Experimental Settings, Hyperparameter Tuning, and Results We trained the text and image feature extractor together. The preprocessed images were resized to 256×256 before passing them through the image feature extractor model. We used the initial learning rate of 10−5 with a linear scheduler. For the BERT model, we used the initial learning rate of 3×10−5 and kept decreasing it using the scheduler. The BERT sequence size was set to 64. A learning rate of 10−5 was used for the fusion module. F1 loss was used as the metric and Adam[24] as the optimizer. The best way to maximize the F1 metric would be to minimize the F1 loss, defined as the 1−F1 score. The issue is that the F1 score is not differentiable. We took inspiration from Lee et Loss Curves 0.62 Textx2-Imagex2 0.60 (Text-Image)x2 0.58 validation loss 0.56 0.54 0.52 0.50 0.48 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 epochs Figure 4: Training Curves for proposed Fusion Modules Table 1 F1 Score analysis for different models Model Name Validation F1 score Bert-base 0.673 EfficientNet-B3 0.617 EfficientNet-B3 + Bert-base (Early Fusion) 0.689 EfficientNet-B3 + Bert-base (Text_X2_Image_X2) 0.725 EfficientNet-B3 + Bert-base (Text_Image_X2) 0.705 Ensemble of Text_X2_Image_X2 and Text_Image_X2 0.712 al.[25] and implemented it for multiclass classification. We accept probabilities instead of actual counts of true positive, true negative, false positive, or false negative. Say the class ”refute” is predicted with probability 0.2, while the true label is ”refute”. Then we calculate true positive as 0.2 and false negative as 0.8. We trained the model using the NVIDIA Tesla P100 GPU with 16 GB of memory and also used RTX5000, having 32 GB of memory. Text_X2_Image_X2 fusion module was trained for 18 epochs, and the model with Text_Image_X2 fusion module was trained for 17 epochs, after which the decrease in validation loss Figure-4 became insignificant. We also tried different architectures, as mentioned in Table 1. We observed that the text- only classifier performs better than the image-only classifier indicating that text provides better insights to news articles. We trained on full data (i.e., text and image combined) us- ing EfficientNetB3 for image and BERT for text which is then followed by an early fusion module [16]( a fusion module that concatenates all the extracted features together and passes them through a Dense layer). Then we tried our proposed models Text_X2_Image_X2 and Text_Image_X2 and found that these outperformed both the early fusion models and the models trained on a single modality. The Text_X2_Image_X2 model gave an F1 score of 0.679 on the test set performing better than the Text_Image_X2 model, which gave a score of 0.664. This was due to the way we concatenate the extracted features. Semantic similarities between similar data(text to text and image to image) lead to a better estimation of differences between claim and reference. We ensembled the models doing a weighted average with different ratios and found an increase in the F1 score to 0.692 on the test set (equal weights). 5. Conclusion and Future Work Combating fake news is crucial as it can destroy one’s credibility and hurt many people. In this paper, we focused on countering the problem of fake news using fact-checking with the help of the FACTIFY dataset. We proposed two novel methods for combining the multimodal features. The Text_X2_Image_X2 method leverages the similarity between the claim and document text features and claim and document image features to predict the required class. The Text_Image_X2 method leverages the similarity between the text and image claim features and text and image document features to predict the required class. Both the proposed methods outperform the existing early fusion network. Our proposed methods showed that combining similar features separately( EfficientNetB3 + Bert-base + Text_X2_Image_X2 and EfficientNetB3 + Bert-base + Text_Image_X2) is more effective than combining all the features together( EfficientNetB3 + Bert-base + Early Fusion ). Future work expects that using deeper CNN models for image feature extractors can improve model performance. Similarly, we can use different variants of BERT for a more robust text feature extractor. Playing with the neural networks used in fusion modules and classifiers can also give better results. 6. Acknowledgement We are indebted to team CyberLabs for their generous support. Also, our work was made possible with the support of JarvisLabs.ai. We thank them for providing a cloud GPU instance for our experiments. We also thank the organisers of DE-FACTIFY 2022 for giving us the opportunity to work on the dataset. References [1] H. Allcott, M. Gentzkow, C. Yu, Trends in the diffusion of misinformation on social media, Research & Politics 6 (2019) 2053168019848554. [2] V. Chenzi, Fake news, social media and xenophobia in south africa, African Identities (2020) 1–20. [3] C. Castillo, M. Mendoza, B. Poblete, Information credibility on twitter, in: Proceedings of the 20th international conference on World wide web, 2011, pp. 675–684. [4] O. Ajao, D. Bhowmik, S. Zargari, Sentiment aware fake news detection on online social networks, in: ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2019, pp. 2507–2511. [5] B. Ghanem, P. Rosso, F. Rangel, An emotional analysis of false information in social media and news articles, ACM Transactions on Internet Technology (TOIT) 20 (2020) 1–18. [6] A. Giachanou, P. Rosso, F. Crestani, Leveraging emotional signals for credibility detec- tion, in: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2019, pp. 877–880. [7] W. Y. Wang, ” liar, liar pants on fire”: A new benchmark dataset for fake news detection, arXiv preprint arXiv:1705.00648 (2017). [8] K. Popat, S. Mukherjee, A. Yates, G. Weikum, Declare: Debunking fake news and false claims using evidence-aware deep learning, arXiv preprint arXiv:1809.06416 (2018). [9] A. Giachanou, E. A. Ríssola, B. Ghanem, F. Crestani, P. Rosso, The role of personality and linguistic patterns in discriminating between fake news spreaders and fact checkers, in: International Conference on Applications of Natural Language to Information Systems, Springer, 2020, pp. 181–192. [10] K. Shu, S. Wang, H. Liu, Understanding user profiles on social media for fake news detection, in: 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), IEEE, 2018, pp. 430–435. [11] S. Singhal, R. R. Shah, T. Chakraborty, P. Kumaraguru, S. Satoh, Spotfake: A multi-modal framework for fake news detection, in: 2019 IEEE fifth international conference on multimedia big data (BigMM), IEEE, 2019, pp. 39–47. [12] Y. Wang, F. Ma, Z. Jin, Y. Yuan, G. Xun, K. Jha, L. Su, J. Gao, Eann: Event adversarial neural networks for multi-modal fake news detection, in: Proceedings of the 24th acm sigkdd international conference on knowledge discovery & data mining, 2018, pp. 849–857. [13] A. Giachanou, G. Zhang, P. Rosso, Multimodal multi-image fake news detection, 2020, pp. 647–654. doi:10.1109/DSAA49011.2020.00091 . [14] N. Vo, K. Lee, Learning from fact-checkers: Analysis and generation of fact-checking language, in: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2019, pp. 335–344. [15] Y. Wang, F. Ma, Z. Jin, Y. Yuan, G. Xun, K. Jha, L. Su, J. Gao, Eann: Event adversarial neural networks for multi-modal fake news detection, 2018, pp. 849–857. doi:10.1145/3219819. 3219903 . [16] I. Gallo, G. Ria, N. Landro, R. L. Grassa, Image and text fusion for upmc food-101 using bert and cnns, in: 2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ), 2020, pp. 1–6. doi:10.1109/IVCNZ51579.2020.9290622 . [17] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv:1810.04805 (2018). [18] K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, arXiv preprint arXiv:1409.1556 (2014). [19] S. Mishra, S. Suryavardan, A. Bhaskar, P. Chopra, A. Reganti, P. Patwa, A. Das, T. Chakraborty, A. Sheth, A. Ekbal, C. Ahuja, Factify: A multi-modal fact verification dataset, in: Proceedings of De-Factify: Workshop on Multimodal Fact Checking and Hate Speech Detection, CEUR, 2022. [20] P. Patwa, S. Mishra, S. Suryavardan, A. Bhaskar, P. Chopra, A. Reganti, A. Das, T. Chakraborty, A. Sheth, A. Ekbal, C. Ahuja, Benchmarking multi-modal entailment for fact verification, in: Proceedings of De-Factify: Workshop on Multimodal Fact Checking and Hate Speech Detection, CEUR, 2022. [21] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, Imagenet: A large-scale hierarchical image database, in: 2009 IEEE conference on computer vision and pattern recognition, Ieee, 2009, pp. 248–255. [22] S. Bird, E. Klein, E. Loper, Natural language processing with Python: analyzing text with the natural language toolkit, ” O’Reilly Media, Inc.”, 2009. [23] A. Buslaev, V. I. Iglovikov, E. Khvedchenya, A. Parinov, M. Druzhinin, A. A. Kalinin, Albumentations: fast and flexible image augmentations, Information 11 (2020) 125. [24] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980 (2014). [25] N. Lee, H. Yang, H. Yoo, A surrogate loss function for optimization of 𝐹𝛽 score in binary classification with imbalanced data, arXiv preprint arXiv:2104.01459 (2021). A. Text Feature Extractor Details The model used was the BASE version of the BERT model having the following parameters: English language uncased, 12 hidden layers (L), 768 hidden sizes (H), 12 self-attention heads (A), 30522 words dictionary (vocab size), 110 million parameters in total. Additionally, a Dropout layer with a probability of 0.3, a Dense layer of size 256, and a Batch-Normalization layer are added to the final layer corresponding to the CLS token.