<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Multimodal Attention is all you need</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Marco</forename><surname>Saioni</surname></persName>
							<email>marco.saioni@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">University G. Marconi</orgName>
								<address>
									<settlement>Rome</settlement>
									<country>IT</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cristina</forename><surname>Giannone</surname></persName>
							<email>c.giannone@unimarconi.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University G. Marconi</orgName>
								<address>
									<settlement>Rome</settlement>
									<country>IT</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Almawave S.p.A</orgName>
								<address>
									<addrLine>Via di Casal Boccone, 188-190 00137</addrLine>
									<settlement>Rome</settlement>
									<country>IT</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Multimodal Attention is all you need</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">BC620F3150DE36EE2D74FDB8B87AC47C</idno>
					<idno type="arXiv">arXiv:2010.11929.</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Transformer</term>
					<term>fake news classification</term>
					<term>multimodal classification</term>
					<term>cross attention</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we present a multimodal model for classifying fake news. The main peculiarity of the proposed model is the cross attention mechanism. Cross-attention is an evolution of the attention mechanism that allows the model to examine intermodal relationships to better understand information from different modalities, enabling it to simultaneously focus on the relevant parts of the data extracted from each. We tested the model using textitMULTI-Fake-DetectiVE data from Evalita 2023. The presented model is particularly effective in both the tasks of classifying fake news and evaluating the intermodal relationship.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Internet has facilitated communication by enabling rapid, immersive information exchanges. However, it is also increasingly used to convey falsehoods, so today, more than ever, the rapid spread of fake news can have severe consequences, from inciting hatred to influencing financial markets or the progress of political elections to endangering world security. For this reason, mitigating the growing spread of fake news on the web has become a significant challenge.</p><p>Fake news manifests itself on the internet through text, images, video, audio, or, in general, a combination of these modalities, which is a multimodal way. In this article, we took the two, text and image, components of news as it proposed, for instance, in a social network. In this work we proposed an approach to automatically and promptly identify fake news. We use the dataset MULTI-Fake-DetectiVE 1 competition, proposed in EVALITA 2023 2 . The competition aims to evaluate the truthfulness of news that combines text and images, an aim expressed through two tasks: the first, which carries out the identification of fake news (Multimodal Fake News Detection); the second, which seeks relationships between the two modalities text and image by observing the presence or absence of correlation or mutual implication (Cross-modal relations in Fake and Real News).</p><p>Our approach proposes a Transformer-based model that focuses on relating the textual and visual embeddings of the input samples (i.e., the vector representations of 1 https://sites.google.com/unipi.it/multi-fake-detective 2 https://www.evalita.it <ref type="bibr" target="#b0">[1]</ref> the text and images it receives as input).</p><p>The aim was to find a way to reconcile the two different representation embeddings because they are learned separately from two different corpora, such as text and images, trying to capture their mutual relationships through some interaction between the respective semantic spaces.</p><p>The remainder of the paper is structured as follows: section 2 presents a brief overview of related work, and section 3 describes the architecture of the proposed model. Section 4 discusses an overview of our experiments. Sections 5 and 6 present the final results and our conclusions, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>The Italian MULTI-Fake-DetectiVE competition <ref type="bibr" target="#b1">[2]</ref> adds to the various datasets and challenges on multimodal fake news recently developed, for instance, Factify <ref type="bibr" target="#b2">[3]</ref> and Fakeddint <ref type="bibr" target="#b4">[4]</ref>. The creation of these competitions shows the interest in this task. The first task of the Italian challenge saw three completely different systems placed on the podium. While the first system POLITO <ref type="bibr" target="#b5">[5]</ref> with a system based on the FND-CLIP multimodal architecture <ref type="bibr" target="#b6">[6]</ref> proposing some ad hoc extensions of CLIP <ref type="bibr" target="#b7">[7]</ref> including sentiment-based text encoding, image transformation in the frequency domain, and data augmentation via back-translation. The Extremita system <ref type="bibr" target="#b8">[8]</ref>, second classified, exploited the LLM capabilities, focusing only on the textual component of each news. They fine-tuned the open-source LMM Camoscio <ref type="bibr" target="#b9">[9]</ref> with the textual part of the dataset. The impressive results show how the textual component plays a primary role in identifying fake news. Despite the significant contribution of the textual component to the task, more and more multimodal approaches are taking hold. In <ref type="bibr" target="#b10">[10]</ref> proposed CNN architecture combining texts and images to classify fake news. In that direction, approaches such as CB-FAKE <ref type="bibr" target="#b11">[11]</ref> incorporate the encoder representations from the BERT model to extract the textual features and comb them with a model to extract the image features. These features are combined to obtain a richer data representation that helps to determine whether the news is fake or real. Visionlanguage models, in general, have gained a lot of interest also in the last years, in the "large models era". Language Vision Models have been proposed during the previous year, with surprising results in many visual language interaction tasks <ref type="bibr" target="#b12">[12]</ref>, <ref type="bibr" target="#b13">[13]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The proposed Model</head><p>The objective was to "engage" specialist models for natural language processing and artificial vision, making them discover and learn bimodal features from text and images collaboratively and harmoniously by applying the teachings of Vaswani et al. <ref type="bibr" target="#b14">[14]</ref>: we decided to follow the path indicated by "Attention is all you need" Vaswani et al. very famous paper, following up on the intuition that the Attention mechanism could provide an important added value to the multimodal model of identification of fake news, becoming a Multimodal Attention (hence the title of this article), i.e. an attention mechanism applied between the two textual and visual modes of news. In fact, while Attention or Self Attention (as described in Vaswani et al. paper) takes as input the embeddings of a single modality and transforms them into more informative embeddings (contextualized embeddings), Multimodal Attention takes as input the embeddings of the two different modalities by combining them and then transforming them into a single embedding capable of capturing any existing relationships between the two input modes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Architecture</head><p>Multimodal Attention is the heart that supports the proposed model, making it capable of exploring the hidden aspects of multimodal communication. As shown at a high level in Figure <ref type="figure" target="#fig_0">1</ref>, the architecture of the proposed model consists of a hierarchical structure with three layers preceded by a pre-processing step. In order, there are: a pre-processing step, an input layer, a cross-modal layer and a fusion layer. It was decided to propose a network that models the consistent information between the two modalities textual and visual starting from State Of The Art pre-trained neural networks. In particular, we use a BERT <ref type="bibr" target="#b15">[15]</ref> pre-trained model to learn the word embeddings by the textual component of news and a ResNet <ref type="bibr" target="#b16">[16]</ref> pre-trained model to learn visual embeddings by the visual component. The two embeddings, belonging to two spaces with different dimensions, are first projected into a uniform, reduced-dimensional space, then related to each other with the strategy of mutual cross-attention to obtain two embeddings subsequently concatenated to provide the input of the last dense classification layer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Pre-processing step</head><p>As a first step it is necessary to process the data made available by the organizers of the MULTI-Fake-DetectiVE competition to produce inputs that are compatible and compliant with those expected from the pre-trained models. The choices made for this preparation or for the pre-processing of the dataset and the data "personalization" strategy will then be described in the following three points:</p><p>• resolution/explosion of 1 : 𝑁 relationships between text and images into 𝑁 times 1 : 1 relationships; • data augmentation with the creation of an additional image to support the original one already present in each example; • management of the textual component, truncated by BERT or rather by the relevant tokenizer to a fixed maximum length of tokens.</p><p>As decided for the visual and textual components, therefore following processing, for each single sample we move from the original pairs &lt; 𝑡, 𝑣 + &gt;, where 𝑣 + indicates the ratio 1 : 𝑁 between text in natural language and images in JPEG format, to the triples appropriately translated into numbers</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>&lt; 𝑡𝑡𝑟𝑢𝑛𝑐, 𝑣, 𝑣𝑎𝑢𝑔 &gt;</head><p>where 𝑡𝑡𝑟𝑢𝑛𝑐 indicates, for each sample, a first-order tensor with 128 values (token), while 𝑣 and 𝑣𝑎𝑢𝑔 denote third-order tensors with (224 × 224 × 3) values (pixels).</p><p>In fact, the first order tensor is the representation of the text in numerical form according to the default strategy of the BERT tokenizer, while the third order tensor is the representation of the images in numerical form according to the RGB coding for ResNet.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Input layer</head><p>This layer receives as input the previously processed dataset, i.e. the text and the images represented in numerical form, passing it to the pre-trained BERT and ResNet models to obtain the respective embeddings, subsequently projected into a space with small and common dimensions to make them comparable and to allow them to collaborate with each other in the subsequent crossmodal layers.</p><p>BERT Encoder Each sample pre-processed and represented in numerical form by the tokenizer is passed as input to the pre-trained BERT model which returns different output tensors for each of them. For the purposes of the classification task object of this study, we consider the pooled_output, a compact representation of all the token sequences given as input to the BERT model, obtained via the special token [CLS]. It is therefore a summary of the information extracted from the entire input dataset whose dimensions evidently depend on the number of hidden units of BERT. Since each text supplied as input to BERT will correspond to a tensor with 768 values real, using vector notation we have that:</p><formula xml:id="formula_0">et = BERT(ttrunc)[𝑝𝑜𝑜𝑙𝑒𝑑_𝑜𝑢𝑡𝑝𝑢𝑡]</formula><p>where et ∈ R ℎ is the word embeddings vector, ttrunc ∈ R 𝑁𝑚𝑎𝑥 is the token input vector and ℎ = 768 is the BERT hidden size. The equation shown refers to a single sample but can be extended to the entire batch of 𝑁 examples processed by BERT. Indicating this batch with Ttrunc ∈ R 𝑁 ×𝑁𝑚𝑎𝑥 , we will have:</p><formula xml:id="formula_1">Et = BERT(Ttrunc)[𝑝𝑜𝑜𝑙𝑒𝑑_𝑜𝑢𝑡𝑝𝑢𝑡]</formula><p>where Et ∈ R 𝑁 ×ℎ is the text embedding matrix learned by the BERT model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ResNet Encoder</head><p>The two images of each sample previously represented in numerical form are passed as input to the pre-trained ResNet model, which returns a visual embedding of size ℎ𝑟 for each example and which represents the features in a compact and semantic form extracted through convolutions and pooling within the ResNet network. In fact, to obtain visual embeddings from a pre-trained neural network like ResNet, we usually take the output of the penultimate layer, i.e. global pooling. In the proposed model, ResNet50V2 was chosen which in global pooling reduces the spatial dimensions of the output tensor to 2048 values and therefore each input image will correspond in output to a vector with ℎ𝑟 = 2048 values, which represents the visual embeddings extracted from the network for that specific image. After obtaining the embeddings for each of the two images, they are concatenated together to obtain a single output tensor which will therefore have size 2 × ℎ𝑟 = 4096. Using the same formalism as the previous text encoder, we have:</p><formula xml:id="formula_2">ev = ResNet(v)[𝑔𝑙𝑜𝑏𝑎𝑙_𝑝𝑜𝑜𝑙𝑖𝑛𝑔]</formula><p>where ev ∈ R ℎ𝑟 is the visual embedding vector and v ∈ R 𝐿×𝐻×𝐶 the input third-order tensor. The indicated equation refers to a single sample but can be extended to the entire batch of 𝑁 examples, therefore indicating the batch with V ∈ R 𝑁 ×𝐿×𝐻×𝐶 , we will have:</p><formula xml:id="formula_3">Ev = ResNet(V)[𝑔𝑙𝑜𝑏𝑎𝑙_𝑝𝑜𝑜𝑙𝑖𝑛𝑔]</formula><p>where Ev ∈ R 𝑁 ×ℎ𝑟 is the visual embedding matrix learned by the ResNet model. Similar discussion for the second image, for which it will be valid at batch level:</p><formula xml:id="formula_4">Ev aug = ResNet(Vaug)[𝑔𝑙𝑜𝑏𝑎𝑙_𝑝𝑜𝑜𝑙𝑖𝑛𝑔]</formula><p>where Ev aug ∈ R 𝑁 ×ℎ𝑟 . By concatenating the two embeddings, we will obtain:</p><formula xml:id="formula_5">Ev ⊕ Ev aug = E concat(v,vaug) ∈ R 𝑁 ×2ℎ𝑟 .</formula><p>From this moment and for simplicity of notation, Ev will refer to E concat(v,vaug) , knowing that this embedding is actually the concatenation of embeddings of an image and the one obtained through random transformations.</p><p>Projection The pre-trained models provide embeddings with different sizes. It is, therefore, necessary to transform them into a space with the same dimensionality to obtain comparable representations. The projection function carries out this task, introduced both to reduce the dimensions of the two embeddings and reduce the computational load, improving the performance of the multimodal model and allowing it to learn more complex patterns. The projection of embeddings is particularly useful in cases where you want to compare the semantic representations of two objects, ensuring that both are aligned in the same reduced semantic space, making them comparable in terms of similarity or distance or facilitating the comparison and analysis of relationships.</p><p>For this model, we selected 𝑑𝑝𝑟𝑗 = 128 as the projection size, reducing both embeddings sizes of the input components.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.3.">Cross-modal layer</head><p>This layer is the heart of the model, which is developed taking inspiration from the behavior of human beings when faced with news made up of text and images. Intuitively, we try to read in the image what is written in the text and to represent in the text what is shown by the image. It can be said that cross-modal attention relations exist between image and text. This is why, to simulate the human process described in a neural model, we relied on the cross attention between the two modalities, a variant of the standard component of multi-head attention capable of capturing global dependencies between text and images.</p><p>In the proposed model, two blocks of crossed attention are activated in the two text-image and image-text perspectives. In the first case, we consider the textual embeddings as queries for the multi-head attention block, while the visual ones as key and value. This should allow the characteristics of the text to guide the model to focus on regions of the image semantically coherent with the text: in fact, if the textual embeddings are considered as queries and the visual ones as key and value, then the attention will be applied to the images in based on compatibility with the text, which is therefore considered the context on which to evaluate the relevance of an image. In this way, attention is focused on the images with respect to how relevant they are to the text, i.e. we try to give importance to the visual features in relation to the context provided by the text. Conversely, in the second case the visual embeddings are the queries, while the keys and values are the textual embeddings, and this should allow the visual features to make the model pay attention to those parts of text consistent with the images. That is, the same thing as in the previous case applies, but the roles between text and image are reversed.</p><p>Wanting to formalize the bidirectional cross-attention between the embeddings of the text E t−projected and those of the images E v−projected , we can write:</p><formula xml:id="formula_6">Ecross−tv = Attention(E t−projected , E v−projected ) Ecross−vt = Attention(E v−projected , E t−projected )</formula><p>where Ecross−tv represents the attention embeddings of image information with respect to the text and Ecross−vt represents attention embeddings of text information compared to images.</p><p>In this layer the dimensions of the embeddings are not modified in any way, therefore we remain in R 𝑁 ×128 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.4.">Fusion layer</head><p>Once you have available the embeddings (textual and visual) learned unimodally in the network, and the crossattention embeddings learned intermodally, it is necessary to implement a fusion strategy that can best balance their respective contributions in the multimodal classification task. Although the architecture of the model would seem to suggest the implementation of the late fusion strategy, it is necessary to observe how the crossattention of the cross-modal layer is already a fusion strategy adopted in the network during learning before the one explicitly implemented in the next fusion layer: this allowed the model to learn shared features during training while maintaining the suitable flexibility between the multimodal components, i.e. without excessively influencing the learning process of each modality separately.</p><p>The concatenation preserves each modality's distinctive features, allowing the model to exploit them during learning, unlike the sum which could lead to the loss of information due to values that can cancel each other out, taking away the model's descriptive capacity. For these reasons, the fusion occurs taking into consideration all four embeddings learned by the model E t−projected , E v−projected , Ecross−tv, Ecross−vt, where the first two provide distinctive unimodal features, while the other two provide correlated and mutually "attentioned" cross-modal features. The hybrid fusion strategy then completes the recipe, providing that pinch of flexibility necessary to give balance to the multimodal classifier. Formally we have the following equation, which aims to make the most of both the information provided by the individual modalities as such, and that provided jointly:</p><formula xml:id="formula_7">E global = (E t−projected ⊕ E v−projected )⊕ Ecross−tv ⊕ Ecross−vt</formula><p>where E global 𝑖𝑛R 𝑁 ×4𝑑 𝑝𝑟𝑗 , where 𝑁 is the size of the batch of examples given as input to the network and 𝑑𝑝𝑟𝑗 = 128.</p><p>The final output of the multimodal model is obtained by applying a densely connected layer with 𝐶 = 4 units and a softmax activation function that returns the probabilities of the four classes. Formally:</p><formula xml:id="formula_8">Y = (E global W + b) O = softmax(Y) with W ∈ R 4𝑑 𝑝𝑟𝑗 ×𝐶 , b ∈ R 1×𝐶 and therefore O ∈ R 𝑁 ×𝐶</formula><p>is a matrix in which each row is a vector with 𝐶 = 4 values representing the conditional (estimated) probability of each class for the relevant sample.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experimental Setup</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Split dataset into training and validation</head><p>To guarantee that the proportions relating to the classes and sources are maintained uniformly in the two sets, the 1034 samples of the dataset are randomly divided following the 80%-20% proportion between training and validation in a stratified way both with respect to the labels, as also happens in the baseline model of the competition MULTI-Fake-DetectiVE and, with respect to the type of source of the news.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Training and validationn</head><p>For our experiment, the model was trained up to 80 epochs with early stopping on using the focal loss <ref type="bibr" target="#b17">[17]</ref> function. It is a dynamically scaled loss cross entropy function, where the scaling factor decays to zero as confidence in the correct class increases. Intuitively, this scaling factor can automatically scale the contribution of easy examples during training and quickly focus the model on difficult examples. For the optimizer we chosed AdamW, given that the models used to analyze text and images were originally pre-trained using this algorithm, which applies weight regularization directly to the model parameters during weight updating, helping to improve the stability and generalization of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Official baseline models</head><p>In the notebook provided by the MULTI-Fake-DetectiVE organizers there is an evaluation strategy on the official dataset which is developed by comparing the performance of the unimodal pre-trained models with a multimodal model: Summary and comparison of the main metrics for the three baseline models on the official dataset.</p><p>by the unimodal textual model, but higher than the score of the unimodal visual model, indicating that the integration of visual and textual information led to an improvement in performance compared to the model visual, but not enough to outperform the text model. This suggests that there may be potential to perform additional optimizations or modality integration strategies to achieve better performance from the multimodal model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Proposed model</head><p>To evaluate the model proposed on the Multimodal Fake News Detection task, we chose to follow the approach used by the organizers in the notebook of the baseline models, i.e. we performed an ablation study on the proposed model: first a unimodal textual model was trained, then a unimodal visual one, then a multimodal one without cross-bi-attention, finally a multimodal one with cross-biattention. The results for the unimodal and multimodal models without cross-bi-attention are in perfect harmony with those of the similar baseline models.</p><p>But the data that catches the eye is that of the accuracy and F1-weighted values of the multimodal model with cross-bi-attention. In particular, its F1-weighted score is almost seven percentage points higher than the proposed textual unimodal model, more than eleven compared to the visual unimodal model and more than nine compared to the multimodal one without cross-bi-attention. Let's see the accuracy and F1-weighted values of the multimodal model proposed with cross-bi-attention against finalist models. Its F1-weighted score is two and a half points higher than that of the winning model of the MULTI-Fake-DetectiVE competition, as evident from the Table <ref type="table" target="#tab_2">3</ref> In fact, if on the one hand the cross-modal layer allows the model to learn multimodal semantics between text and images, the fusion layer enhances it by improving its stability, capacity and performance thanks to the skip connection which provides the gradient with a useful direct path during backpropagation to flow without tending to zero, bringing significant and additional information into each layer of the network.</p><p>All the results described up to this point are obtained by measuring the model on the Multimodal Fake News Detection task of the competition covered by this work. As mentioned, the organizers also proposed a second task Cross-modal relations in Fake and Real News, aimed at verifying the robustness of the model to changing tasks without any human intervention. a clear improvement in performance in solving the task even compared to the winning model of the competition. This is a very important result, because it demonstrates the network's ability to adapt to changes in tasks and changes in training data, which is not at all a given.</p><p>The data preparation strategy in the Pre-processing step provides the model with more information to learn from, the real strength can be identified in the Cross-modal Layer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions</head><p>The Internet has facilitated the multimodality of communication by enabling rapid information exchanges that are increasingly immersive but increasingly used to convey falsehoods. In this study, a multimodal model for identifying fake news was proposed which is based on the mechanism of cross attention between the representations of the features learned by the network on the textual component of the news and those learned on the visual component associated with it.</p><p>Many multimodal models are based on the concatenation of features learned from distinct modalities which, despite having good performance, however, limit the potential of the interaction between the features themselves.</p><p>From the experiments carried out, the use of crossattention demonstrated significant improvements in the performance of the model proposed in this work compared to the first two models classified in the MULTI-Fake-DetectiVE competition for both tasks requested by the organizers, despite the dataset available for training is very small in size and unbalanced both with respect to the categories to be predicted and with respect to the source of the news. Despite the intrinsic complexity of the two tasks, the cross-layer of the proposed model manages to express the representations learned from the text and images of a news story in a harmonious, collaborative and synergistic way, balancing their contributions and preventing one from taking over the other.</p><p>Future developments concern the components of the model which could use a Visual Transformer <ref type="bibr" target="#b18">[18]</ref> instead of the ResNet in order to relate textual embeddings and visuals both generated by training a Transformer network.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Proposed model architecture.</figDesc><graphic coords="2,333.12,84.19,142.36,247.35" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell>• Text-only model: model trained only on textual</cell></row><row><cell>features, extracted with a pre-trained BERT net-</cell></row><row><cell>work.</cell></row><row><cell>• Image-only model: model trained only on the</cell></row><row><cell>visual features of images, extracted with a pre-</cell></row><row><cell>trained ResNet18 network.</cell></row><row><cell>• Multi-modal model: model trained on the con-</cell></row><row><cell>catenation of text and image features, extracted</cell></row><row><cell>separately with the previous two only-model.</cell></row><row><cell>The F1-weighted score values of the three baseline mod-</cell></row><row><cell>els are shown in Table 1. The textual model is therefore</cell></row><row><cell>the most effective among the three baseline models in</cell></row><row><cell>classifying fake news and the visual one has lower per-</cell></row><row><cell>formance than the textual model. The multimodal model</cell></row><row><cell>obtained an F1-weighted score lower than that obtained</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>Table 2 reports the respective accuracy and F1-weighted values.</figDesc><table><row><cell>Model</cell><cell cols="2">Accuracy F1-weighted</cell></row><row><cell>Proposed Multi-modal ⊗</cell><cell>0.541</cell><cell>0.537</cell></row><row><cell>Proposed Text-only</cell><cell>0.472</cell><cell>0.469</cell></row><row><cell>Proposed Multi-modal ⊕</cell><cell>0.460</cell><cell>0.445</cell></row><row><cell>Proposed Image-only</cell><cell>0.418</cell><cell>0.422</cell></row><row><cell>Table 2</cell><cell></cell><cell></cell></row><row><cell cols="3">Ablation study on the proposed model: accuracy and F1-</cell></row><row><cell cols="3">weighted. The ⊗ symbol indicates cross-bi-attention enabled,</cell></row><row><cell cols="3">while ⊕ indicates cross-bi-attention disabled (i.e. concatena-</cell></row><row><cell>tion enabled).</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>. As supposed and hoped, the mechanism Final comparison between all the analyzed models and the proposed model.of crossed attention seen from the two text-image and image-text perspectives enriched by the skip connection provided by the simple concatenation of the two different embeddings, provides the model with that extra edge that allows it to dig background in the relationships between textual and visual features. By combining bilateral cross-attention and residual connection, tasks of the cross-modal layer and the fusion layer respectively, significant semantic and semiotic interrelations are obtained in favor of the performance of the classifier which becomes more precise and sensitive.</figDesc><table><row><cell>Model</cell><cell cols="2">Accuracy F1-weighted</cell></row><row><cell>Proposed Multi-modal</cell><cell>0.541</cell><cell>0.537</cell></row><row><cell>PoliTo -FND-CLIP-ITA</cell><cell>-</cell><cell>0.512</cell></row><row><cell>ExtremITA -Suede_LoRA</cell><cell>-</cell><cell>0.507</cell></row><row><cell>Baseline Multi-modal</cell><cell>0.480</cell><cell>0.442</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>shows the accuracy and F1-weighted values for the proposed model called to express itself on the Cross-modal relations task, together with the baseline and winner models of the MULTI-competition Fake-DetectiVE. The results show</figDesc><table><row><cell>Model</cell><cell cols="2">Accuracy F1-weighted</cell></row><row><cell>Proposed Multi-modal</cell><cell>0.529</cell><cell>0.527</cell></row><row><cell>PoliTo -FND-CLIP-ITA</cell><cell>-</cell><cell>0.517</cell></row><row><cell>Baseline Multi-modal</cell><cell>-</cell><cell>0.442</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Result summary on Task 2.</figDesc><table /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<ptr target="https://ceur-ws.org/Vol-3473" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Lai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Menini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Russo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Sprugnoli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Venturi</surname></persName>
		</editor>
		<meeting>the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)<address><addrLine>Parma, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 7th-8th, 2023. 2023</date>
			<biblScope unit="volume">3473</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Multi-fake-detective at evalita 2023: Overview of the multimodal fake news detection and verification task</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondielli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dell'oglio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Marcelloni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Passaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sabbatini</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3473/paper32.pdf" />
	</analytic>
	<monogr>
		<title level="j">CEUR WORKSHOP PROCEEDINGS</title>
		<imprint>
			<biblScope unit="volume">3473</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">S</forename><surname>Suryavardan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Patwa</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Factify 2: A multimodal fake news and satire news dataset</title>
		<author>
			<persName><forename type="first">M</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Reganti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chadha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Sheth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chinnakotla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ekbal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar</surname></persName>
		</author>
		<ptr target="http://dblp.uni-trier.de/db/conf/defactify/defactify2023.html#SuryavardanMPCR23" />
	</analytic>
	<monogr>
		<title level="m">DE-FACTIFY@AAAI</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Sheth</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Ekbal</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3555</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Fakeddit: A new multimodal benchmark dataset for fine-grained fake news detection</title>
		<author>
			<persName><forename type="first">K</forename><surname>Nakamura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Wang</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.lrec-1.755" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twelfth Language Resources and Evaluation Conference, European Language Resources Association</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Béchet</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Blache</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choukri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Cieri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Declerck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Goggi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Isahara</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mazo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Moreno</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Twelfth Language Resources and Evaluation Conference, European Language Resources Association<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="6149" to="6157" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Polito at multi-fake-detective: Improving FND-CLIP for multimodal italian fake news detection</title>
		<author>
			<persName><forename type="first">L</forename><surname>D'amico</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Napolitano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vaiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cagliero</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3473/paper35.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Lai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Menini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Russo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Sprugnoli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Venturi</surname></persName>
		</editor>
		<meeting>the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)<address><addrLine>Parma, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 7th-8th, 2023. 2023</date>
			<biblScope unit="volume">3473</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Multimodal fake news detection via clip-guided learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2205.14304.arXiv:2205.14304" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Learning transferable visual models from natural language supervision</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hallacy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Goh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Krueger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.00020</idno>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Extremita at EVALITA 2023: Multi-task sustainable scaling to large language models at its extreme</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Hromei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Basili</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3473/paper13.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Lai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Menini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Russo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Sprugnoli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Venturi</surname></persName>
		</editor>
		<meeting>the Eighth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2023)<address><addrLine>Parma, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 7th-8th, 2023. 2023</date>
			<biblScope unit="volume">3473</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Camoscio: an italian instruction-tuned llama</title>
		<author>
			<persName><forename type="first">A</forename><surname>Santilli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rodolà</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2307.16456.arXiv:2307.16456" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Multimodal fake news detection</title>
		<author>
			<persName><forename type="first">I</forename><surname>Segura-Bedmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Alonso-Bartolome</surname></persName>
		</author>
		<ptr target="https://www.mdpi.com/2078-2489/13/6/284" />
	</analytic>
	<monogr>
		<title level="j">Information</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Cb-fake: A multimodal deep learning framework for automatic fake news detection using capsule neural network and bert</title>
		<author>
			<persName><forename type="first">B</forename><surname>Palani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Elango</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename></persName>
		</author>
		<idno type="DOI">10.1007/s11042-021-11782-3</idno>
	</analytic>
	<monogr>
		<title level="j">Multimedia Tools and Applications</title>
		<imprint>
			<biblScope unit="volume">81</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Cogvlm: Visual expert for pretrained language models</title>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Hong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2311.03079.arXiv:2311.03079" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Improved baselines with visual instruction tuning</title>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">J</forename><surname>Lee</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2310.03744.arXiv:2310.03744" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.03762</idno>
		<title level="m">Attention is all you need</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Deep residual learning for image recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR.2016.90</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<imprint>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="770" to="778" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Focal loss for dense object detection</title>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dollár</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1708.02002</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">An image is worth 16x16 words: Transformers for image recognition at scale</title>
		<author>
			<persName><forename type="first">A</forename><surname>Dosovitskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Beyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kolesnikov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Weissenborn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Unterthiner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dehghani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Minderer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Heigold</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Houlsby</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
