<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ULMFiT f or Twitter Fake News Spreader Profiling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>H. L. Shashirekha</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>F. Balouchzahi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore - 574199</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <abstract>
        <p>21st century is named as the age of information technologies. Social applications such as Facebook, Twitter, Instagram, etc. have become a quick and huge media for spreading news over the internet. At the same time, the ability for the wide spread of news that is of low quality with intentionally false information is creating havocs causing damage to the extent of losing lives in the society. Such news is termed as fake news and detecting the fake news spreader is drawing more attention these days as fake news can manipulate communities' minds and also social trust. Until date, many studies have been done in this area and most of them are based on Machine Learning and Deep Learning approaches. In this paper, we have proposed a Universal Language Model Fine-Tuning model based on Transfer Learning to detect potential fake news spreaders on Twitter. The proposed model collects wiki text data to train the Language Model to capture general features of the language and this knowledge is transferred to build a classifier using fake news spreaders dataset provided by PAN 2020 to identify the fake news spreader. The results obtained on PAN 2020 fake news dataset are encouraging.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        In this era, social media is overwhelming the lives of people and people are sharing
various information using different platforms of social media such as Google+,
Facebook, WhatsApp and Twitter [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The velocity of news spreading on internet is
highly increasing due to the availability of various social media platforms and pocket
friendly mobile data packs. Social media has become more attractive especially for
the younger generation mainly because of the inherent benefits of fast dissemination
of information and easy access to the information [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. At the same time, the ability for
the wide spread of news that is of low quality with intentionally false information is
creating havocs causing damage to the extent of losing lives in the society [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>Two major concepts of fake news are veracity and intention. Veracity is about the
news that includes some information and the authenticity of that content is possible to
be verified as they are. For example, in case of a news about earthquake in Japan, the
probability of this news being true is higher but it is a challenge to prove that it is fake
or not. Intention refers to the goal of spreader to use false information intentionally to
mislead the reader.</p>
      <p>
        Fake news is not a new challenge as people have been exposed to propaganda,
tabloid news, and satirical reporting since ages. But nowadays, the heavy dependence
on the internet, trending stories on social media, new methods of monetizing content,
etc., have been found to rely on information without using trustworthy traditional
media outlets [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Fake news is hazardous since it is spread to manipulate readers’
opinions and beliefs [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Hence, detecting fake news spreaders becomes very much
important in today’s scenario and is gaining popularity day by day as users play a key
role in creating and sharing incorrect or false information intentionally or accidently
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In spite of many systems including automatic detection systems and human based
systems, detection of fake news spreaders is still a challenging task [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Detecting fake news spreaders in Twitter can be modeled as a typical binary Text
Classification (TC) problem that labels a given news spreader as fake or genuine. TC
is a Supervised Machine Learning (ML) technique that automatically assigns a label
from the predefined set of labels to a given unlabelled input. It has wide applications
in various domains, such as target marketing, medical diagnosis, news classification,
and document organization [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. There are several popular approaches for TC in
general and for fake news spreader profiling in particular. In this paper, we propose a
Universal Language Model Fine-Tuning (ULMFiT) model for fake news spreader
detection based on Transfer Learning (TL).
      </p>
      <sec id="sec-1-1">
        <title>1.1 Transfer Learning</title>
        <p>
          TL is generally known as one of the novel inventions in the field of Deep Learning
and Computer Vision. Conventionally, in ML every model is built from the scratch
using a specific dataset. However, a model based on TL approach uses the knowledge
obtained from building one model called as a source model in building another model
called as target model. The former model is called as source task and later the target
task. While the source task uses one dataset called as source dataset to build/learn the
source learning system or source model, target task uses the knowledge obtained in
building the source model along with the target dataset used for fine tuning the target
model. For example, the source model can be a Language model (LM) that represents
the general features of a language, target model can be TC, source dataset can be
Wikipedia text and the target dataset can be fake news [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. LM is a probability
distribution over word sequences in a language and introduces a useful hypothesis
space for many other NLP tasks [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. As the knowledge obtained in building the
source model is transferred to build the target model, learning is named as Transfer
Learning. Figure 1 illustrates the difference between conventional ML and TL. After
the introduction of TL, LM has drawn more attention as it acts as an informative
knowledge of a language.
        </p>
      </sec>
      <sec id="sec-1-2">
        <title>1.2 ULMFiT</title>
        <p>
          ULMFiT is a model based on TL and can be used for many NLP tasks such as TC
and NER [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. It uses the knowledge of LM as source model and then fine tunes the
target model using the task-specific data or target dataset. Figure 2 represents
architecture of ULMFiT. It includes 3 steps i) pre-training LM using large corpus like
Wikipedia to capture the high-level language features and the resultant model is
called as pre-trained LM ii) fine-tune the target model using pre-trained LM and
taskspecific or target dataset iii) final model which accepts the test/unlabelled data to
assign a label.
        </p>
        <p>
          The advantage of TL is, when a given dataset is too small to train a learning model
the knowledge obtained in a pre-trained LM on a source dataset can be transferred to
the target task, resulting in the improvement of the target model even when the source
and target datasets have different distributions or features [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ][
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>The rest of the paper is organized as follows. Section 2 gives the related work
followed by the proposed methodology in section 3. While section 4 describes the
experiments and results, section 5 gives the conclusion of the paper.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2 Related Works</title>
      <p>In spite of the availability of many automated tools and techniques for the
detection of fake news spreaders, it is still a challenging task. Some of the relevant
works are mentioned below:</p>
      <p>
        An Artificial Neural Network model for Language Identification task for Indian
native Languages namely Tamil, Hindi, Kannada, Malayalam, Bengali and Telugu
written in Roman script has been explored by Hamada et. al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The data sets used in
task are collection of comments from different regional newspapers and Facebook
pages. They obtained an accuracy score of 35.30 %. The same authors also obtained
accuracies of 47.60% and 47.30% respectively in another work using ensemble
classifier made of multinomial Bayes, SVM and random forest tree [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Francisco et.
al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] proposed Low Dimensionality Representation (LDR) for language variety
identification and has applied LDR to the age and gender identification task at the
PAN Lab at CLEF. The results they obtained are competitive with the best
performing teams in the author profiling task.
      </p>
      <p>
        Shu et. al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] constructs a real-world dataset by measuring users trust level of
"experienced1" and "native2" users on fake news. They have performed a comparative
analysis of explicit and implicit profile features between these user groups, which
reveals their potential to differentiate fake news. Shu et. al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] have explored the fake
news problem from a data mining perspective, including feature extraction and model
construction and have reviewed different approaches for fake news detection. Bilal et
al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] presents an approach based on a combination of emotional information from
documents using a deep learning network. The authors used one dataset including
trusted news (real news) created from English Gig word corpus and another dataset is
a collection of news from seven different unreliable news sites as false news and have
reported an F1 score of 96%. A Bot detection approach using behavioral and other
informal cues is proposed by Andrew et. al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. They have used random forest
classifier and a gradient boosting classifier and also applied a hyper parameter
optimization on over 476 million revisions that has been collected from Wikipedia
articles. They have reported the model performance as 88% precision and 60% recall.
      </p>
      <p>
        EmoCred model based on LSTM neural network proposed by Anastasia et. al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]
incorporates emotional signals to differentiate between credible and non-credible
claims. It accepts word embeddings as input from claims and a vector of emotional
signals. The authors used Politifact3 that contain the text of the claims, the speaker,
and the credit rating of each claim. Six different credibility ratings: true, mostly true,
half true, mostly false, false, and pants-on-fire has been combined into two classes as
true and false and obtained 61.7% F1 score for generating the emotional signals.
“DeClarE” is an automated end-to-end neural network model proposed by Kashyap
et. al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. They capture signals from external evidence articles and model joint
interactions between various factors like the context of a claim, the language of
reporting articles, and the trustworthiness of their sources. Their model was evaluated
on Snopes4, Politifact 5, and a SemEval Twitter rumor dataset and obtained F1 scores
of 79% and 68% for Snopes and Politifact respectively and a macro accuracy score of
57% for SemEval dataset.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>
        An overview of the proposed fake news spreader detection model is described in
Figure 3. The model constructed using the state-of-the-art ULMFiT architecture
developed by Howard et. al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] consists of pre-training the LM and then fine-tuning
the fake news spreader detection model by using the pre-trained LM and fake news
spreader dataset provided by PAN2020. Two separate models are constructed to
detect the fake news given in English and Spanish. Inspired by Stephen et. al. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ],
      </p>
      <sec id="sec-3-1">
        <title>1 Users who are able to recognize fake news items like false</title>
        <p>2 Users who are more likely to believe fake news
3 It is a fact-checking website where the credibility of different claims is investigated.
4 www.snopes.com
5 www.politifact.com
LM and Target classifier are created using text.models module from fastai library.
This module implements the encoder for an ASGD Weight-Dropped LSTM
(AWDLSTM) which can be plugged in with a decoder to create an LM and also with some
classifying layers to create a text classifier.</p>
        <p>
          AWD-LSTM is a regular LSTM to which several regularization and optimization
techniques are applied and built layer by layer by grabbing a PyTorch neural network
model [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. Its architecture as described by Howard and Ruder [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] consists of a word
embedding of size 400, 3 layers and 1150 hidden activations per layer. The
AWDLSTM has been dominating the state-of-the-art language modeling and many studies
on word-level models incorporate AWD-LSTMs. It also has shown noticeable results
on character-level models [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
        </p>
        <p>LM also called as source learning model is trained on the source data collected
from English/Spanish Wikipedia. Source data set usually is an unannotated data set
that contains general domain texts to train LM to obtain general features like grammar
of the language. A sufficiently large English/Spanish text data are collected from
Wikipedia to create a source dataset of English/Spanish language respectively and
LM is trained to learn the general features of the language. Wikipedia articles that
were available in the month of January 2020 are collected in xml format and then the
sentences are extracted from the raw text using WikiExtractor6 module. Once the
source model completes its learning the knowledge thus learned is used to build the
target task of fake news spreader detection. The knowledge can also be saved for
future use for other English/Spanish NLP applications. Details of source dataset for
both the languages are given in Table 1.
3.2</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Target Model</title>
      <p>
        The target model is created using the knowledge obtained from LM followed by
fine-tuning the model using the target dataset. The pre-trained LM is used to train
target task data for various cycles to fine-tune the knowledge based on target task.
Target dataset is the labeled data used for classification tasks which is provided by
PAN for registered users only. The dataset consists of 300 XML files in a folder per
language (English, Spanish) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. Each folder contains:
      </p>
      <sec id="sec-4-1">
        <title>6 https://github.com/attardi/wikiextractor</title>
        <p> An XML file per author (Twitter user) consisting of 100 tweets each and the
name of the XML file corresponds to the unique author id.</p>
        <p> A truth.txt file with the list of authors and ground truth.</p>
        <p>
          The details of the dataset provided by PAN are given in Table 2. Target data is
preprocessed and then used for fine-tuning the classification task. Preprocessing
involves tokenization, removing punctuations and stop words, lemmatization and
removing other unwanted characters. Emojis are small images used to express
emotion and are useful in text analysis [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Hence, they are converted to respective
words or phrases and those words or phrases are treated similar to content bearing
words.
        </p>
        <p>Language</p>
        <sec id="sec-4-1-1">
          <title>English</title>
        </sec>
        <sec id="sec-4-1-2">
          <title>Spanish</title>
          <p>
            As per PAN 2020 rules for submitting software in Virtual Machine (VM), learning
model has to be first constructed locally and saved followed by loading the model in
PAN VM and finally submitting the model through TIRA Integrated Research
Architecture submission system [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ]. ULMFiT model is created using Google Colab7
as it requires GPU and higher RAM size in learning cycles.
          </p>
          <p>The proposed model was evaluated through PAN submission system and the
performance of model was made available by the task moderator. Model's runtime
reported by PAN is 00:35:48 (hh:mm:ss). Almost half of this time is spent on loading
the model using fastai library and rest for predictions. Details of results obtained by
the proposed model are given in Table 3. The proposed model resulted with 64%
accuracy for Spanish and 62% for English language data.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>7 https://colab.research.google.com/</title>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6 Conclusion</title>
      <p>This paper presents ULMFiT model for profiling fake tweet spreaders based on
Transfer Learning approach. The proposed model is initially trained on a general
domain English/Spanish data collected from Wikipedia to build an LM model, and
then the acquired knowledge is transferred to build the fake news spreader detection
task as the target model. The model resulted with 64% accuracy for Spanish and 62%
for English language data. Further, the data collected from Wikipedia and LM can be
used for any other English/Spanish NLP task.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Nayel</given-names>
            <surname>Hamada</surname>
          </string-name>
          <article-title>A., and</article-title>
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          . “Mangalore University INLI@ FIRE2018:
          <article-title>Artificial Neural Network and Ensemble based Models for INLI”</article-title>
          .
          <source>In FIRE (Working Notes)</source>
          , pp.
          <fpage>110</fpage>
          -
          <lpage>118</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Shu</given-names>
            <surname>Kai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Suhang</given-names>
            <surname>Wang</surname>
          </string-name>
          , and Huan Liu. “
          <article-title>Understanding User Profiles on Social Media for Fake News Detection”</article-title>
          .
          <source>In 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR)</source>
          , pp.
          <fpage>430</fpage>
          -
          <lpage>435</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Shu</given-names>
            <surname>Kai</surname>
          </string-name>
          , Amy Sliva, Suhang Wang,
          <string-name>
            <given-names>Jiliang</given-names>
            <surname>Tang</surname>
          </string-name>
          , and Huan Liu. “
          <article-title>Fake News Detection on Social Media: A Data Mining Perspective”</article-title>
          .
          <source>ACM SIGKDD Explorations Newsletter 19, No. 1</source>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>36</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Haber</given-names>
            <surname>Morey</surname>
          </string-name>
          . “
          <article-title>The Real Risks of Fake News”</article-title>
          .
          <source>Risk Management</source>
          <volume>64</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>10</fpage>
          -
          <lpage>12</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Ghanem</given-names>
            <surname>Bilal</surname>
          </string-name>
          , Paolo Rosso, and Francisco Rangel. “
          <article-title>An Emotional Analysis of False Information in Social Media and News Articles”</article-title>
          .
          <source>ACM Transactions on Internet Technology (TOIT) 20, no. 2</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Giachanou</given-names>
            <surname>Anastasia</surname>
          </string-name>
          , Esteban A.
          <string-name>
            <surname>Ríssola</surname>
          </string-name>
          , Bilal Ghanem, Fabio Crestani, and Paolo Rosso. “
          <article-title>The Role of Personality and Linguistic Patterns in Discriminating Between Fake News Spreaders and Fact Checkers”</article-title>
          .
          <source>In International Conference on Applications of Natural Language to Information Systems</source>
          , Springer, Cham, pp.
          <fpage>181</fpage>
          -
          <lpage>192</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Vo</given-names>
            <surname>Nguyen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Kyumin</given-names>
            <surname>Lee</surname>
          </string-name>
          .
          <article-title>“Learning from Fact-checkers: Analysis and Generation of Fact-checking Language”</article-title>
          .
          <source>In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pp.
          <fpage>335</fpage>
          -
          <lpage>344</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Aggarwal Charu</surname>
            <given-names>C.</given-names>
          </string-name>
          , and Cheng Xiang Zhai.
          <article-title>“A Survey of Text Classification Algorithms</article-title>
          .
          <source>In Mining Text Data”</source>
          . Springer pp.
          <fpage>163</fpage>
          -
          <lpage>222</lpage>
          , Boston, MA,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Faltl</given-names>
            <surname>Sandra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Schimpke</surname>
          </string-name>
          and
          <string-name>
            <given-names>Constantin</given-names>
            <surname>Hackober</surname>
          </string-name>
          . “Ulmfit:
          <article-title>State-OfThe-Art in Text Analysis”</article-title>
          .
          <source>Seminar Information Systems (WS18/19)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Howard</surname>
            <given-names>Jeremy</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>Sebastian</given-names>
            <surname>Ruder</surname>
          </string-name>
          . “
          <article-title>Universal Language Model Fine-Tuning for Text Classification”</article-title>
          , arXiv preprint arXiv:
          <year>1801</year>
          .06146,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Semwal</surname>
            <given-names>Tushar</given-names>
          </string-name>
          , Promod Yenigalla, Gaurav Mathur, and
          <string-name>
            <surname>Shivashankar</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Nair</surname>
          </string-name>
          .
          <article-title>“A Practitioners Guide to Transfer Learning for Text Classification Using Convolution Neural Networks”</article-title>
          .
          <source>In Proceedings of the 2018 Society for Industrial and Applied Mathematics (SIAM) International Conference on Data Mining</source>
          , pp.
          <fpage>513</fpage>
          -
          <lpage>521</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. Pan Sinno Jialin, James T. Kwok, and Qiang Yang, “
          <article-title>Transfer Learning via Dimensionality Reduction”</article-title>
          ,
          <string-name>
            <surname>In</surname>
            <given-names>AAAI</given-names>
          </string-name>
          , vol.
          <volume>8</volume>
          , pp.
          <fpage>677</fpage>
          -
          <lpage>682</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Nayel</surname>
            , Hamada A., and
            <given-names>H. L.</given-names>
          </string-name>
          <string-name>
            <surname>Shashirekha</surname>
          </string-name>
          . “Mangalore-University@
          <article-title>INLI-FIRE2017: Indian Native Language Identification using Support Vector Machines and Ensemble approach”</article-title>
          .
          <source>In FIRE (Working Notes)</source>
          , pp.
          <fpage>106</fpage>
          -
          <lpage>109</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14. Rangel Francisco, Marc Franco-Salvador, and Paolo Rosso.
          <article-title>“A Low Dimensionality Representation for Language Variety Identification”</article-title>
          .
          <source>In International Conference on Intelligent Text Processing and Computational Linguistics</source>
          , Springer, Cham, pp.
          <fpage>156</fpage>
          -
          <lpage>169</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Hall</surname>
            <given-names>Andrew</given-names>
          </string-name>
          , Loren Terveen, and Aaron Halfaker. “
          <article-title>Bot Detection on Wikidata Using Behavioral and Other Informal Cues”</article-title>
          .
          <source>Proceedings of the ACM on Human-Computer Interaction</source>
          , Vol.
          <volume>2</volume>
          , No.
          <source>CSCW, Article</source>
          <volume>64</volume>
          ,
          <year>November 2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Giachanou</surname>
            <given-names>Anastasia</given-names>
          </string-name>
          , Paolo Rosso, and Fabio Crestani. “
          <article-title>Leveraging Emotional Signals for Credibility Detection”</article-title>
          .
          <source>In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pp.
          <fpage>877</fpage>
          -
          <lpage>880</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Popat</surname>
            <given-names>Kashyap</given-names>
          </string-name>
          , Subhabrata Mukherjee,
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Yates</surname>
          </string-name>
          , and Gerhard Weikum. “
          <article-title>DeClarE: Debunking Fake News and False Claims Using Evidence-Aware Deep Learning”</article-title>
          . arXiv preprint arXiv:
          <year>1809</year>
          .06416,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Merity</surname>
            <given-names>Stephen</given-names>
          </string-name>
          , Nitish Shirish Keskar, and Richard Socher. “
          <article-title>Regularizing and Optimizing LSTM Language Models”</article-title>
          .
          <source>arXiv preprint arXiv: 1708.02182</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Rangel</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giachanou</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghanem</surname>
            <given-names>B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Rosso P.</surname>
          </string-name>
          “
          <article-title>Overview of the 8th Author Profiling Task at PAN 2020: Profiling Fake News Spreaders on Twitter”</article-title>
          . In: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Eickhoff</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>A</surname>
          </string-name>
          . Névéol (eds.)
          <article-title>CLEF 2020 Labs and Workshops, Notebook Papers</article-title>
          ,
          <source>CEUR Workshop Proceedings</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Potthast</surname>
            <given-names>Martin</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tim Gollub</surname>
          </string-name>
          , Matti Wiegmann, and Benno Stein.
          <article-title>“TIRA Integrated Research Architecture”</article-title>
          .
          <source>In Information Retrieval Evaluation in a Changing World</source>
          , Springer, Cham, pp.
          <fpage>123</fpage>
          -
          <lpage>160</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>