<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>GDUF_DM at FakeDeS 2021: Spanish Fake News Detection with BERT and Sample Memory</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Xixuan Huang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jieying Xiong</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shengyi Jiang</string-name>
          <email>jiangshengyi@163.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Guangzhou Key Laboratory of Multilingual Intelligent Processing, Guangdong University of Foreign Studies</institution>
          ,
          <addr-line>Guangzhou</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Information Science and Technology, Guangdong University of Foreign Studies</institution>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Fake news widely spread on the Internet has had a negative impact on society. This article reports the solution of Spanish fake news detection purposed by our team GDUFS_DM in IberLEF 2021 shared task. Our purpose is to use BERT and Sample Memory with an attention mechanism to detect Spanish fake news. To capture richer semantic information in long news texts, we used BERT to encode the news headline and the news beginning and end part to keep more information instead of using a simple truncation strategy. In addition, we also use a matrix parameter initialized by sample representation (we call it Sample Memory), combine with the attention mechanism, our model can capture the relationship information between samples which strengthens the model's robustness in the inference stage. Our submission result achieved the first place on the leaderboard, which fully reflects the advantages of our model.</p>
      </abstract>
      <kwd-group>
        <kwd>Fake News Detection</kwd>
        <kwd>Spanish</kwd>
        <kwd>BERT</kwd>
        <kwd>Sample Memory</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Fake news refers to the news articles that are intentionally and verifiably false [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The
rapid development of online news media platforms not only provided convenience for
readers to obtain news information, but also provided soil for the breeding and
dissemination of fake news. The publication of fake news is often intentional, and some
individuals or organizations may publish different types of fake news for different
purposes. Fake news can not only be used to insult and slander individuals, but it can also
disrupt social order, instigate political unrest, or even undermine the peace and stability
of the international community. What's worse, researches on the dissemination of fake
news shows that fake news is significantly faster, deeper, and wider distributed than
true news [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Therefore, it’s has important practical significance to know how to use
the machine to automatically and accurately identify false news.
      </p>
      <p>
        The FakeDeS@IberLEF 2021 [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ] provided us with a fake news detection corpus
in Mexican Spanish [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The corpus mainly collects Mexican Spanish fake news from
websites and contains quite balanced data of real and fake news on 9 different topics,
which is intended to encourage people to actively research the identification of fake
news in Mexican Spanish to solve the problem of detecting fake news articles in
Mexican Spanish spread through digital media. The distribution of the dataset is shown in
Table 1. Our team GDUFS_DM also participated in this evaluation and achieved first
place on the leaderboard. In this report, we will review our solution for this task,
namely, Mexican Spanish Fake News Detection with BERT and Sample Memory (see
Section 3.3 for details).
Numerous scholars had conducted extensive research in text features and emotional
features to improve the effect of fake news detection. Ajao et al. pointed out that there
is a relationship between the news veracity and the sentiment of the published text and
attached a sentiment feature (ratio of the number of negative and positive words) to
help the plain text fake news detectors [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Instead of attaching a unique feature, Zhang
et al. verified the difference between dual emotions in fake news and real news, and
proposed a dual emotion feature to represent dual emotions and the relationship
between them for fake news detection [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Przybyap concluded that the writing style of
fake news has certain characteristics, so they designed two new classifiers: a neural
network classifier and a model classifier based on stylometric feature [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Wang et al.
proposed an enhanced weakly supervised fake news detection framework, WeFEND,
which can utilize user reports as weak supervision to expand the amount of training
data for fake news detection, given the dynamic nature of news and the reality that
labeled samples may become outdated quickly [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Yi Xie et al. proposed a fake news
detection framework to make full use of characterize users by utilizing social user
graphs [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        However, the majority of studies on automated fake news detection have been
limited to English documents, and few have evaluated works in other languages. Moreover,
the spread of deceptive news tends to be a worldwide problem, so we need to study
fake news not only in English, but also look at the world and detect fake news in other
languages. Some scholars had also studied fake news detection of some low-resource
languages. Nankai Lin et al. proposed the CharCNN-RoBERTa model to detect fake
news in the Urdu Language [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Hugo Queiroz Abonizio et al. evaluated textual
features not linked to a specific language when describing textual data for detecting news
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. News corpora written in American English, Brazilian Portuguese and Spanish
were explored to investigate complexity, stylometric and psychological textual
features. As regards the Mexican Spanish, the MEX-A3T@IberLEF2020 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] has called
methods for aggressiveness and fake news detection in Spanish in Mexico. Samuel
Arce-Cardenas et al. evaluated the combination of basic text classification techniques,
including six machine learning algorithms, two methods for extracting keywords, and
two preprocessing techniques [14]. The best results they ran showed an F1-macro score
of 0.815 for fake news. Esaú Villatoro-Tello el at. [15] evaluated Supervised
Autoencoder (SAE) learning algorithms in a text classification task. They used three different
sets of features as input, namely classical word n-grams, char n-grams and Spanish
BERT encodings, and obtained the best performance ( = 85.66%) in the fake news
classification task.
3
3.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <sec id="sec-2-1">
        <title>Overview</title>
        <p>The model we finally proposed in this task is shown in Fig. 1. The beginning and end
part of the news text is feed into BERT for obtaining two text embeddings (which were
called Head Embedding and Tail Embedding). Then after an element-wise addition was
applied in those two embeddings, we calculated the dot-product attention between the
result and Sample Memory utils to obtain Memory Embedding. Finally, the Beginning
Embedding, End Embedding, and Memory Embedding are stitched together to
calculate the output result. The specific components will be explained in detail below. In
addition, we also use tricks such as gradient accumulating, early stop, and hierarchical
learning rates.
It has been shown that the use of pre-trained language models (PTMs) significantly
improves the performance of text classification, and also reduces the amount of labeled
sample data required in supervised learning [16]. In this evaluation task, we also used
one of the representatives of the pre-trained model, BERT [17] (Bidirectional Encoder
Representations from Transformers). In the pre-training section, the model needs to
learn the general semantic information of the language from a large-scale unlabeled
corpus, according to the pre-training task we set. In fine-tuning section, we can use it
as a feature capture model in downstream tasks to obtain the embedding of text or
tokens and use different finetune frameworks according to the specific task with labeled
data. One disadvantage of the pre-trained model is that pre-training often takes a lot of
computing power and time which may be a difficult thing for us. Fortunately, DCC
Canete J et al. released the Spanish BERT model on an open-source platform called
Transformers [18,19]. This model has two versions of cased and uncased for us to use
which were also the main BERT models used in this evaluation.</p>
        <p>Most pre-trained models set the maximum sequence length to 512, which is not very
friendly to long texts such as news texts. A common solution is truncation, that is, only
a sequence of tokens of a limited length is retained. Previous research has found that
for text classification, keeping the head and tail tokens at the same time can achieve
better results than keeping only the head tokens or only the tail tokens, which means
that the head and tail parts of the text may contain more information than the middle
part [20]. We concatenated news headlines and news texts in the training set and
validation set. After tokenization, we calculated the length distribution of the token
sequence (see Fig. 2). The average sequence length reached 546, and nearly 41% of the
sequence length is greater than 512. This means that if only a simple truncation method
is used, a lot of useful information may be lost. To solve this problem, we adopted a
simple method, respectively taking the first 512 tokens and the last 512 tokens for
encoding, and concatenating them to get the embedding of the news text to retain as much
information as possible.</p>
        <p>450
400
means that the news in the test corpus may contain topics which are not part of the
training corpus, and some test data may be inconsistent with the training corpus in terms
of language style. In order to improve the robustness and generalization ability of the
model, we also designed a network structure which called "Sample Memory". Sample
Memory contains m vectors with the same dimensions as Head Embedding and Tail
Embedding (we call it Memory Utils). Before network training, the Head Embedding
and Tail Embedding of m samples need to be used for the initialization of Sample
Memory. The model accepts a news and applies Dot-product Attention between sample
embedding and Sample Memory utils to obtain Memory Embedding. The calculation
formula of Memory Embedding is shown in formula 1.</p>
        <p>∗ [ 1 ,  2 , … ,   ] ∗ [ 1,  2, … ,   ]


(1)
 
Where  
= 
is Memory Embedding,    
refers to the element-wise addition
result of Head Embedding and Tail Embedding, and  1,  2, … ,   respectively refers
to m Utils in Sample Memory. Therefore, even if the news text received by the model
is very different from the training set during inferencing, Memory Embedding at this
time is still not going to change much, which means that the experience learned by the
model in the past is still going to be useful. In addition, we concatenate Memory
Embedding with Head Embedding and Tail Embedding in the next network layer as usual,
the model therefore can better learn the semantic information of the news text itself.
3.4</p>
      </sec>
      <sec id="sec-2-2">
        <title>Tricks</title>
        <p>Gradient Accumulation. Limited by the memory size of the training GPU device, the
batch size of our model during training can only reach 4. Researchers have found that
increasing the batch size appropriately will help the model loss decrease more stably
[21]. Therefore, the effect of batch size of 32 is approached by accumulating the
gradient of every 8 training steps during model training.</p>
        <p>Early stop. Early stop is a Widely used trick in deep learning to avoid overfitting. After
each training epoch is finished, we run an evaluation on the validation dataset to obtain
validation loss. If the loss value does not continue to decrease within 3 epochs, then the
model training will be stopped and the model with the best performance on the
validation set will be taken as the model to be submitted.</p>
        <p>Differential learning rates. It has been shown that the features captured by different
layers in the neural network may be different [22]. Howard and Ruder pointed out that
different layers in the neural network should be set to different learning rates according
to specific tasks, and the experiment shows that setting different learning rates
according to different layers is beneficial to the model to achieve better results [23]. This
inspired us to set different learning rates for the parameters of the BERT model, the
parameters of the Sample Memory and the parameters of the output layer during the
model training process.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiment</title>
      <p>In our experiment, we only used the string concatenated by the news title and body as
input. The final result we submitted used the bert-base-spanish-wwm-cased model [15]
published by DCC UChile and the spanberta-base model published by Skim AI
Technologies [24]. We used 1e-5, 1e-4, and 1e-3 initial learning rates for BERT parameters,
Sample Memory parameters, and output layer parameters. Before model training, we
selected all training samples, validation samples, and test samples to initialize Sample
Memory. In fact, we also tried some machine learning models (using TFIDF for feature
extraction) for comparison, such as Support Vector Machine (SVM), Naive Bayes
(NB), Logistic Regression (LR), Decision Tree (DT), Gradient Boosting Decision Tree
(GBDT) and Random Forest (RF). Table 2 shows the results in validation and test set
that was reported in accuracy score, precision score for fake (Fake-P), recall score for
fake(Fake-R) and f1-score for fake(Fake-F1).</p>
      <p>Model
SVM
NB
LR
DT
GBDT
RF
Ours(spanberta-base-cased)
Ours(bert-base-spanish-wwm-cased)
Second best system (in leaderboard)</p>
      <p>In order to explore why Sample Memory works in our model, we also designed the
following experiment on the validation set: initialize the Sample Memory randomly
instead of using news corpus for initialization, and simply remove Sample Memory (see
Table 3). It can be seen that even randomly initialized Sample Memory can also
improve the performance while using real news samples to initialize Sample Memory can
achieve better results. We also found something interesting that the classification
precision of fake news has improved after removing Sample Memory, although the recall
score has decreased significantly. This may indicate that Sample Memory can guide the
model to seek "reference objects" from past samples to guide decision-making.
Although errors may sometimes occur, fake news can be more easily detected thanks to
Sample Memory.
In this report, we presented the solution of team GDUFS_DM in the IberLEF 2021
shared task to detect fake news in Spanish. We proposed to use BERT to encode the
beginning and end of the news text, and then applied the Sample Memory module let
the model to learn the relationship between samples. We also applied some training
tricks gradient accumulation, early stop and differential learning rates. The results show
that our model got the first place in the ranking. In our future work, we will consider
using information such as URLs and topics added to the news to enhance the
performance of fake news detection.</p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgements</title>
      <p>This work was supported by the Key Field Project for Universities of Guangdong
Province (No. 2019KZDZX1016), the National Natural Science Foundation of China (No.
61572145) and the National Social Science Foundation of China (No. 17CTQ045). The
authors would like to thank the anonymous reviewers for their valuable comments and
suggestions.
14. Arce-Cardenasa S, Fajardo-Delgadoa D, Álvarez-Carmonab M Á.: TecNM at MEX-A3T
2020: Fake News and Aggressiveness Analysis in Mexican Spanish. In: Proceedings of the
Iberian Languages Evaluation Forum (IberLEF 2020) co-located with 36th Conference of
the Spanish Society for Natural Language Processing (SEPLN 2020). pp. 265-272. CEUR
Workshop Proceedings (2020).
15. Ramírez-de-la-Rosa G, Parida S, Kumar S, et al.: Idiap and UAM Participation at
MEXA3T Evaluation Campaign. In: Proceedings of the Iberian Languages Evaluation Forum
(IberLEF 2020) co-located with 36th Conference of the Spanish Society for Natural
Language Processing (SEPLN 2020). pp. 252-257. CEUR Workshop Proceedings (2020).
16. Peters M E, Neumann M, Iyyer M, et al.: Deep contextualized word representations. arXiv
preprint arXiv:1802.05365 (2018).
17. Devlin J, Chang M W, Lee K, et al.: Bert: Pre-training of deep bidirectional transformers for
language understanding. arXiv preprint arXiv:1810.04805 (2018).
18. Canete J, Chaperon G, Fuentes R, et al.: Spanish pre-trained bert model and evaluation data.</p>
      <p>PML4DC at ICLR (2020).
19. BETO: Spanish BERT, https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased,
last accessed 2021/06/01
20. Sun C, Qiu X, Xu Y, et al.: How to Fine-Tune BERT for Text Classification?. arXiv e-prints
arXiv: 1905.05583 (2019).
21. You Y, Gitman I, Ginsburg B.: Scaling sgd batch size to 32k for imagenet training. arXiv
preprint arXiv:1708.03888 (2017).
22. Yosinski J, Clune J, Bengio Y, et al.: How transferable are features in deep neural networks?.</p>
      <p>arXiv preprint arXiv:1411.1792 (2014).
23. Howard J, Ruder S.: Universal language model fine-tuning for text classification. arXiv
preprint arXiv:1801.06146 (2018).
24. Spanish Bert pretrained model released by Skim AI Technologies,
https://huggingface.co/skimai/spanberta-base-cased, last accessed 2021/06/01</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Shu</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sliva</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            <given-names>S</given-names>
          </string-name>
          , et al.:
          <article-title>Fake news detection on social media: A data mining perspective</article-title>
          .
          <source>ACM SIGKDD explorations newsletter 19(1)</source>
          ,
          <fpage>22</fpage>
          -
          <lpage>36</lpage>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Vosoughi</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roy</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aral</surname>
            <given-names>S.:</given-names>
          </string-name>
          <article-title>The spread of true and false news online</article-title>
          .
          <source>Science</source>
          <volume>359</volume>
          (
          <issue>6380</issue>
          ),
          <fpage>1146</fpage>
          -
          <lpage>1151</lpage>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gómez-Adorno</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Posadas-Durán</surname>
            <given-names>J P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bel-Enguix</surname>
            <given-names>G</given-names>
          </string-name>
          , and
          <article-title>Clau-dia P.: Overview of fakedes task at iberlef 2021: Fake news detection in spanish</article-title>
          .
          <source>Procesamiento del Lenguaje Natural</source>
          ,
          <volume>67</volume>
          (
          <issue>0</issue>
          ), (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Manuel</given-names>
            <surname>Montes</surname>
          </string-name>
          , Paolo Rosso, Julio Gonzalo, Ezra Aragón, Rodrigo Agerri,
          <string-name>
            <surname>Miguel Ángel</surname>
          </string-name>
          Álvarez-Carmona, Elena Álvarez Mellado, Jorge
          <string-name>
            <surname>Carrillo-de-Albornoz</surname>
          </string-name>
          , Luis Chiruzzo, Larissa Freitas, Helena Gómez Adorno, Yoan Gutiérrez, Salud María Jiménez Zafra, Salvador Lima,
          <string-name>
            <surname>Flor Miriam</surname>
          </string-name>
          Plaza-de-Arco and Mariona Taulé (eds.):
          <source>Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2021</year>
          ), CEUR Workshop Proceedings, (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Posadas-Durán</surname>
            <given-names>J P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez-Adorno</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorov</surname>
            <given-names>G</given-names>
          </string-name>
          , et al.:
          <article-title>Detection of fake news in a new corpus for the Spanish language</article-title>
          .
          <source>Journal of Intelligent &amp; Fuzzy Systems</source>
          ,
          <year>2019</year>
          ,
          <volume>36</volume>
          (
          <issue>5</issue>
          ):
          <fpage>4869</fpage>
          -
          <lpage>4876</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ajao</surname>
            <given-names>O</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhowmik</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zargari</surname>
            <given-names>S.:</given-names>
          </string-name>
          <article-title>Sentiment aware fake news detection on online social networks</article-title>
          .
          <source>In: ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          . IEEE, pp.
          <fpage>2507</fpage>
          -
          <lpage>2511</lpage>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Zhang</surname>
            <given-names>X</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            <given-names>X</given-names>
          </string-name>
          , et al.:
          <article-title>Mining Dual Emotion for Fake News Detection</article-title>
          . arXiv e-prints arXiv:
          <year>1903</year>
          .
          <volume>01728</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Przybyla</surname>
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Capturing the Style of Fake News</article-title>
          .
          <source>In: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          . vol.
          <volume>34</volume>
          (
          <issue>01</issue>
          ), pp.
          <fpage>490</fpage>
          -
          <lpage>497</lpage>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Wang</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ma</surname>
            <given-names>F</given-names>
          </string-name>
          , et al.:
          <article-title>Weak supervision for fake news detection via reinforcement learning</article-title>
          .
          <source>In: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          . vol.
          <volume>34</volume>
          (
          <issue>01</issue>
          ), pp.
          <fpage>516</fpage>
          -
          <lpage>523</lpage>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Xie</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            <given-names>X</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xie</surname>
            <given-names>X</given-names>
          </string-name>
          , et al.:
          <article-title>A Fake News Detection Framework Using Social User Graph</article-title>
          .
          <source>In: Proceedings of the 2020 2nd International Conference on Big Data Engineering</source>
          . pp.
          <fpage>55</fpage>
          -
          <lpage>61</lpage>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Lina</surname>
            <given-names>N</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fua</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jianga</surname>
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Fake News Detection in the Urdu Language using CharCNNRoBERTa. In: The 12th edition of the Forum for Information Retrieval Evaluation (FIRE</article-title>
          <year>2020</year>
          ). pp.
          <fpage>447</fpage>
          -
          <lpage>451</lpage>
          . (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Abonizio H Q</surname>
            ,
            <given-names>de Morais</given-names>
          </string-name>
          J I,
          <string-name>
            <surname>Tavares</surname>
            <given-names>G M</given-names>
          </string-name>
          , et al.:
          <article-title>Language-independent fake news detection: English, Portuguese, and Spanish mutual features</article-title>
          .
          <source>Future Internet</source>
          <volume>12</volume>
          (
          <issue>5</issue>
          ),
          <volume>87</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Aragón</surname>
            <given-names>M E</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jarquín</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gómez M M</surname>
          </string-name>
          , et al.:
          <article-title>Overview of mex-a3t at iberlef 2020: Fake news and aggressiveness analysis in mexican Spanish</article-title>
          .
          <source>In: Notebook Papers of 2nd SEPLN Workshop on Iberian Languages Evaluation Forum (IberLEF)</source>
          , Malaga,
          <string-name>
            <surname>Spain.</surname>
          </string-name>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>