<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SAMS: Human-in-the-loop Approach to Combat the Sharing of Digital Misinformation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shaban Shabani</string-name>
          <email>shaban.shabani@unibas.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zarina Charlesworth</string-name>
          <email>zarina.charlesworth@he-arc.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Sokhn</string-name>
          <email>maria.sokhn@hes-so.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Heiko Schuldt</string-name>
          <email>heiko.schuldt@unibas.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Mathematics and Computer Science, University of Basel</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institut de Digitalisation, University of Applied Sciences Western Switzerland (HES-SO)</institution>
          ,
          <addr-line>Neuchâtel</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Spread of online misinformation is an ubiquitous problem especially in the context of social media. In addition to the impact on global health caused by the current COVID-19 pandemic, the spread of related misinformation poses an additional health threat. Detecting and controlling the spread of misinformation using algorithmic methods is a challenging task. Relying on human fact-checking experts is the most reliable approach, however, it does not scale with the volume and speed with which digital misinformation is being produced and disseminated. In this paper, we present the SAMS Human-in-the-loop (SAMS-HITL) approach to combat the detection and the spread of digital misinformation. SAMS-HITL leverages the fact-checking skills of humans by providing feedback on news stories about the source, author, message, and spelling. The SAMS features are jointly integrated into a machine learning pipeline for detecting misinformation. First results indicate that SAMS features have a marked impact on the classification as it improves accuracy by up to 7.1%. The SAMSHITL approach goes one step further than the traditional human-in-the-loop models in that it helps raising awareness about digital misinformation by allowing users to become self fact-checkers.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Digital Misinformation</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Crowdsourcing</kwd>
        <kwd>Human-in-the-Loop</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Advances in mobile technology have allowed for an unprecedented spread of information and both
mis- and disinformation. The ease of transmission and sharing, the use of social media and
messaging apps coupled with the increasing penetration of the Internet, provides a fertile ground for its
spread [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. As pointed out by Ciampaglia [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], the risk is the massive, uncontrolled, and often systemic
spread of untrustworthy content.
      </p>
      <p>
        Digital misinformation comes in a variety of forms from entirely false to the integration of one or
two misleading sentences in a piece of real news or just a provocative misleading title in introduction
to a correct piece of news. In addition to this, one also finds rumor, hoaxes, satire, and conspiracy
theories contributing to what can be characterized as an online false information ecosystem [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. One
can group such information under the umbrella term of misinformation. More recently, one also sees
a distinction between misinformation, which can be spread with or without intent to mislead, and
disinformation, which intends to spread false information [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        The problem with misinformation is that it is pervasive and runs through all types of media from
print to radio to online. The latter grew considerably during 2016 American presidential election and
with the onset of the COVID-19 pandemic in March of 2020 has now taken on alarming proportions.
In the words of T. A. Ghebreyesus1, director-general of the WHO, speaking of COVID-19: “We are
not just fighting an epidemic; we are fighting an infodemic” . Digital misinformation spreads faster and
more easily than this virus, and is just as dangerous. Unlike the virus, however, COVID-19 related
news has two strains true and false with the latter inundating social media channels, and going largely
unverified. The expression infodemic was first used in 2003 by Rothkopf [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] when writing about the
SARS epidemic and highlighting the negative impact that misinformation had on controlling the then
health crisis – a crisis far from the size of what we are now experiencing with COVID-19. In today’s
COVID-19 influenced world, we are indeed dealing with an infodemic and the question is how best to
control it and fight the spread of misinformation. In order to counter this, and in light of the high level
of user distrust towards online fact-checking services, there is an urgent need to train individuals to
evaluate the veracity of the information they are receiving and sharing, and give them the tools to
become fact-checkers in their own right.
      </p>
      <p>
        In recent years, fact-checking services have become the norm for journalists and are now also easily
accessible by the public. Although the majority address issues in the political arena, SNOPES2, a well
known service, started out primarily by debunking urban legends. Such services certainly have a
role to play in response to the challenge of online misinformation [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], however, there is increasing
interest in coming up with an automated and scalable response [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Despite the availability of
factchecking services, research suggests that there is a high level of distrust for such services [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. Yet
another argument in support of the development of an individual user application.
      </p>
      <p>
        This research focuses specifically on digital misinformation. As this is technology-based, the
solutions considered are also often only technology-related [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Some advocate that the response to
the online spread of misinformation is through technology [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and the integration of Artificial
Intelligence (AI), others lean more towards human fact checkers. An alternate possibility is through
a combination of the two allowing for the development of a high level performing model used by
individuals. Research in the area of mixed-initiative fact-checking [
        <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
        ] suggests that AI alone
cannot be as accurate as when the human element is integrated in the fact-checking process.
Humanin-the-loop AI (HAI) systems face diferent challenges in terms of efective performance due to fact
that individuals are involved. It is, however, possible to train models to a significant level of accuracy.
      </p>
      <p>
        We suggest going one step further in the battle to slow-the-flow by taking an HAI approach as well
as involving those who are at the source by getting them on board as fact-checkers in their own rights.
In order to do this, we have developed a user friendly tool that both identifies the veracity of the news
and calls on the user to self-check four critical indicators on their own. Our proposed framework
is called SAMS. Following a review of the published research and literature on credibility indicators
[
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ] and fact-checking guides [
        <xref ref-type="bibr" rid="ref17 ref9">9, 17</xref>
        ] the choice of a limited number of checks seemed to be
appropriate.
      </p>
      <p>
        In order to be best armed to counter the spread of misinformation, it is important to see who is
spreading it. Spring [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] suggests grouping the spreaders into seven categories ranging from the
“Joker” to the “Politician” and including “well-intentioned family members”. Zannettou et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] go
one step further including even bots for a total of ten categories. Keeping in mind the fact that we
were looking for a limited number of indicators and yet ones that could be applied to all categories,
the ones that repeatedly came to the top of the list were: source, author, message, and spelling (SAMS).
1https://www.who.int/dg/speeches/detail/munich-security-conference
2https://www.snopes.com/fact-check/
We went with these and launched into the development of a prototype.
      </p>
      <p>In this paper, we propose and evaluate the SAMS-HITL method. It uses supervised machine
learning models to classify news articles into false and true. It leverages the fact-checking skills of humans
by providing answers to SAMS related questions of the news articles. The human feedback about the
four SAMS indicators is joined with the automatically extracted features from the text of news articles.
We evaluate this approach on a recently published dataset with articles related to COVID-19.
Preliminary results show that the SAMS human-in-the-loop (SAMS-HITL) approach outperforms methods
that rely only on automated techniques. Feeding the model with information about the source, author,
message, and spelling provides higher accuracy in the classification task.</p>
      <p>Our contributions can be summarized as follows:
• We conceptualized SAMS-HITL as a user friendly option to check information, which calls on
human input and is backed up by machine learning models
• We designed a crowdsourcing task that leverages the human cognitive skills of online crowd
workers on providing answers to SAMS questions, using an aggregation method to infer the
true answer based on multiple judgments
• We implemented and evaluated SAMS-HITL approach on a dataset with COVID-19 related
news articles. The proposed technique performs better than automatic classification models.
Preliminary results indicate that SAMS features have a marked impact on the classification as
it improves accuracy by up to 7.1%.</p>
      <p>The remainder of the paper is structured as follows: Section 2 introduces the concept behind SAMS
and describes the individual components. Section 3 introduces the dataset and describes the
implementation. Section 4 details the evaluation and results of our methods. Section 5 presents related
work and Section 6 concludes with suggestions for further research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Concept</title>
      <p>In this section, we present the components of the SAMS-HITL approach. Figure 1 illustrates the
overall architecture of SAMS.</p>
      <sec id="sec-2-1">
        <title>2.1. Machine Learning Component</title>
        <p>
          Considering the rapid growth of online data and spread of misinformation, and the high impact it has
on the society, eficient and efective data processing tools are essential. Approaches based on machine
learning and deep learning techniques [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] have been comprehensively considered for fake news
detection. The core component of SAMS is the supervised machine learning model which analyzes
the news content. This model consists of two phases: i) feature extraction, and ii) model construction.
        </p>
        <p>Feature extraction is performed on the text coming from the headline and the body text. The
headline of a news item is a short text that is meant to catch the attention of the reader, whereas the body
text is the main part that details the news story. We consider two types of features: statistical and
sentiment features based on linguistic characteristics. The statistical features are extracted using the
Term-Frequency Inverse-Document-Frequency (tf-idf) algorithm which measures the importance of
words in the text document. Analyzing the sentiment of the news is very important especially when
taking into account that much of the misinformation being spread started out as disinformation with
the intent to deceive rather than to report objectively, sometimes for political or financial gain, and
in the COVID-19 pandemic situation to exploit public fear and uncertainty. Therefore, the sentiment
features focus on capturing the objectivity aspect, the mood, modality, and polarity of the reported
news. Additionally, the length of news stories is an important aspect as misinformation generated on
social network channels tends to be short and catchy.</p>
        <p>Model construction builds the machine learning model to perform the classification, in order to
diferentiate between false and true news. For the evaluations, we selected four diferent
state-of-theart algorithms: Logistic Regression (LOG), Random Forest (RF), Gradient Boosting Classifier (GBC),
and Support Vector Machines (SVM).</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. SAMS – Source, Author, Message, Spelling</title>
        <p>
          Classifying news articles by relying solely on machine learning models based on news content is a
challenging task. One reason is that spreaders of misinformation have advanced their writing style
and the language used in the news with the aim of distorting the truth and bypass detection by
stylebased models. Another very important factor is the length of the news: false stories, especially in
the era of the COVID-19 pandemic, tend to be short and alerting, making it dificult to automatically
analyze the message conveyed by such stories. Research in the area of mixed-initiative fact-checking
[
          <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
          ] suggests that machine learning alone cannot be as accurate as when the human element
is integrated in the fact-checking process. We have identified that important features when
performing fact-checking are: source, author, message, and spelling. Answering the questions about the four
features of SAMS automatically is non-viable. In contrast, humans have the potential to do better, as
they can perform fact-checking skills, by searching for facts on trustworthy data resources. We define
a process with tips and tricks to easily answer each of the questions for SAMS.
        </p>
        <p>Source – taking a critical look at the source, both data and metadata, is the first step. The goal is
to understand if the news stories have sources and if the sources are reliable. To better evaluate if the
sources are trustworthy, we take a look at where the information originated, inspect if the references
are stated and if so, trace the references checking if they are correct and trustworthy.</p>
        <p>Author – in principle, real and serious news articles always have an author. Therefore, the first
step is to identify if there is an author of the news item. If so, further inspections include if the author
is a journalist, their afiliation, academic or professional credentials. Furthermore, a check for related
publications by the same author can be made.</p>
        <p>Message – the message should be clear, balanced, and unbiased. Guidelines to identify
misinformation suggest checking for unsupported or outrageous claims, if there is a push to share the
information, lack of quotes, references or contributing sources, and identifying if the headlines provoke
strong emotions.</p>
        <p>Spelling – reputable sources will proofread material prior to publishing. Misinformation tends to
have grammar mistakes such as repeated spelling mistakes, poor grammar, incorrect punctuation, use
of diferent fonts, the writing of entire words or phrases written in capital letters, etc.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Human-in-the-Loop Approach</title>
        <p>Obtaining reliable information related to the four aspects of SAMS is crucial for our approach towards
misinformation detection. While experts such as journalists are well trained to search for the right
data sources to find facts, employing them becomes expensive. Considering the amount and velocity
of potential disseminated digital misinformation, this approach is not scalable in terms of time.</p>
        <p>On the other hand, crowdsourcing has been widely deployed for small tasks as it leverages the
collective human skills of extensive online crowd worker communities. In specific scenarios,
crowdsourcing has shown to be an alternative service to replace experts with specific domain knowledge
for labelling. In this work, we design a crowdsourcing component which aggregates the inputs from
multiple users to infer the true answers related to SAMS questions. The output of the crowd answers
is a vector of four binary values, each value corresponding to the SAMS questions. As could be seen in
Figure 1, the output is encoded into a binary feature vector which later is appended to the feature
vector generated using the tf-idf algorithm and the sentiment features described in Section 2.1. Finally,
the concatenated feature vector is used for training and evaluating the machine learning models.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset and Implementation</title>
      <p>In this section, after presenting the dataset we used to evaluate our approach, we describe the
implementation of the SAMS pipeline.</p>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>
          In this experimental setup, we use the CoAID dataset collected and annotated by Cui and Lee [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. The
dataset consists of true and false news about COVID-19 from diverse sources mainly covering websites
and social network platforms. There are several types of entries such as “news articles” collected from
fact-checking reliable sources, “claims” posted by oficial channels of WHO, “user engagement” which
include twitter posts and replies, and other “social platform posts” such as Facebook, Instagram, etc.
For each entry, there is the title, abstract, content, keywords, and URL of the article or the post. Our
interest is in analyzing news posts that contain potentially longer text and posted on various social
media channels (online newspapers, blogs, communication apps etc.), therefore we focus only news
articles, skipping the twitter posts with short text. As a result, we filtered the false and true news
from the CoAID dataset, ending up with 1,127 true samples, and 266 false samples. Considering that
the two classes are imbalanced, finally from the dataset we selected all 266 available false entries and
264 randomly sampled real news articles. Figure 2 illustrates the distribution of news length by word
count and Table 1 describes the descriptive statistics of the articles length. Average length of true and
false news articles is 35 and 33 words, respectively.
        </p>
        <p>Class
false
true</p>
        <p>Min</p>
        <p>Max</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Implementation</title>
        <p>The first step towards implementing a model is data pre-processing. Since this is a text classification
task, text cleaning is a useful process and that includes removing frequent words that provide
nonunique information to the model (stop words) and special characters, and applying stemming and
lemmatization. This part is important for extracting features using the tf-idf technique. On the other
hand, extracting sentiment features is possible on the raw text and this is done with the pattern.en tool
[21]. The sentiment features include: i) polarity which is given as a value between -1.0 (completely
negative) and +1.0 (completely positive); ii) subjectivity which is a value between 0.0 (very objective)
and 1.0 (very subjective); iii) modality feature represents the degree of certainty as a value between
-1.0 and +1.0, where values higher than +0.5 represent facts ; iv) mood feature is a categorical value
based on auxiliary verbs and the answer can be either “indicative”, “imperative”, “conditional”, or
“subjunctive”. Additionally, we add the word count to the feature vector.</p>
        <p>The aim of this work is evaluating the SAMS-HITL approach and the importance of the four
indicators in the classification task. Doing that, called for having answers to the four SAMS questions for
every news record from the dataset. Initially, we labelled manually the 530 dataset entries. A trained
annotator used an in-house developed web annotation interface to answer the SAMS questions for
each dataset entry. The annotation was a tedious task that took approximately 30 hours. After that,
we designed a crowdsourcing job on the Microworkers3 crowdsourcing platform. Each news story
was used to generate a Human Intelligence Task (HIT), asking online crowd participants to provide
the answers to SAMS questions which were stated as follows:
1. Is there a source in this news article? – Yes/No.
2. Is there an author in this news article? – Yes/No.
3. Is the message of this news article clear, unbiased, and balanced? – Yes/No.</p>
        <p>4. Is spelling correct on this news article? – Yes/No.</p>
        <p>A HIT contained the URL of the original news story, the headline, and the body text. Crowd workers
were instructed to click on the news link, inspect and analyze the article always considering the four
questions that they were asked to answer. We used data quality control mechanisms [22] such as
redundancy where for each task we asked three workers from three diferent regions: USA, Europe,
and Asia. Analysis on demographics and dynamics of crowd workers on crowdsourcing platforms has
shown that these regions are mostly represented [23] and this is the case in the Microworkers platform
as well. Additionally, collecting judgments from crowd workers from diferent regions could be an
important factor as diversity matters when it comes to label quality [24]. Furthermore, task design
techniques are important element for high quality data, therefore guidelines with tips and examples
were part of the instructions in the crowdsourced job.</p>
        <p>The choice of obtaining multiple judgments for the same question from diferent user demographics
can increase the quality of crowdsourced data. Depending on the task complexity, various aggregation
techniques can be applied, such as voting strategies, profile-based or iterative aggregation algorithms
[25]. The majority voting is the simplest as it is non-iterative and does not require pre-processing,
aggregating each object independently by choosing the label with highest votes. In our scenario, for a
single HIT that has SAMS questions, we would get three responses for each of the four questions and
the aggregation selects the answers with highest votes. Since crowd workers can have diferent levels
of expertise, profile-based strategies take into account information from their past contributions to
build a ranking score, as well incorporate additional information such as location, domain of interests
etc. The reputation score of crowd workers can be updated dynamically based on their performance
on the existing task. Iterative algorithms are based on a sequence of computational rounds where in
each round the probabilities of possible labels for each object are computed and updated repeatedly
until convergence. For the answer aggregation in our SAMS-HITL approach, we applied the Dawid
and Skene algorithm [26], a model that is based on the Expectation-Maximization (EM) principle to
model the worker’s reliability with a confusion matrix.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>In this section, we outline the evaluation of the proposed SAMS-HITL approach. We used a 10-fold
cross validation for evaluating the performance of the models, and accuracy and f1 score as evaluation
metrics. To analyze the evaluation of models and the importance of the three diferent sets of features
described in Sections 2.1 and 2.2, we run the evaluation of the models for each setting separately. We
consider the following combinations of features:
(i) tf-idf features (TF)
(ii) tf-idf + sentiment features (TFS)
(iii) tf-idf + sentiment features + SAMS trained annotator features (TFST)
(iv) tf-idf + sentiment features + SAMS crowd features (TFSC)</p>
    </sec>
    <sec id="sec-5">
      <title>5. Related Work</title>
      <p>
        In recent years, automatic misinformation classification has been extensively used under specific
supervised scenarios. Shu et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] explore the characterization of fake news in social media and the
data mining perspectives. As an emerging topic, misinformaton has drawn attention of the research
communities in diferent disciplines. As a result, several datasets have been published, related to
political news [27], rumor debunking [28], fake vs. satire detection [29], FEVER dataset for verification
of textual sources [30], and a more recent dataset with news related to the COVID-19 pandemic [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
Significant eforts have been made on exploring the potential of deep learning [27, 31] and the
linguistic and semantic aspects [32, 33].
      </p>
      <p>Crowdsourcing as a methodology can assist in classification of news articles by fact-checking the
statements. Tschiatschek et al. [34] propose a strategy of flagging news considered as false by
social network users, and deploys an aggregation method to select subset of those flagged news for
evaluation by experts. A recent work by Roitero et al. [35] analyzed how the crowd users assessed
the truthfulness of false and true statements related to COVID-19 pandemic. Results show that the
crowd is able to accurately classify statements and achieve a certain level of agreement with expert
judgments. In response to combating the COVID-19 misinformation, Li et al. [36] developed the
Jennifer chatbot which helps users to easily find information related to COVID-19. The chatbot provides
reliable sources and diverse topics maintained by a global group of volunteers.</p>
      <p>
        Considering the sensitivity and the risk of misinformation spreading on one hand, and the
limitations of both automated and human-based methods, hybrid human-machine approaches have been
envisioned [
        <xref ref-type="bibr" rid="ref12">12, 37</xref>
        ]. For instance, the hybrid machine-crowd [38] approach has demonstrated higher
accuracy for classification of fake and satiric stories. It uses a high-confidence switching method
where crowd feedback is requested whenever the ensemble of machine learning models fails to achieve
unanimity and high accuracy. A hybrid human-machine interactive approach [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] based on a
probabilistic graphical model combines machine learning model predictions with crowd annotations for
fact-checking. The follow-up user study [39] shows that predictions from automated models can help
users in assessing claims correctly, hence tending to trust the system, even when model predictions
were wrong. However, enabling interaction and having transparent model predictions has the
potential of training the users to build their own skills for fact-checking.
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and Future Work</title>
      <p>In this paper, we have addressed the issue of classifying news stories related to the COVID-19
pandemic. We presented SAMS-HITL framework for misinformation detection. SAMS-HITL combines
statistical and sentiment based features automatically extracted from the text of the news articles with
the features related to the source, message, author, and spelling of the article obtained via
crowdsourcing. Preliminary results showed that the four SAMS features are the most important features of the
classification model and it has high impact on the overall classification accuracy. In summary, our
proposed framework leverages the eficiency of machine learning models over a large amount of data and
the quality of human intelligence for fact-checking. This method is helpful for social networks which
could benefit from the high availability of their platform members and leverage their fact-checking
skills to provide feedback on SAMS questions on news articles that are posted and shared on their
platform. The SAMS-HITL approach goes one step further than the traditional HAI models in that it
calls on the users to answer the four questions themselves thus raising user awareness about digital
misinformation. The SAMS-HITL prototype application is currently being developed. Our objective
is to help users check news articles, time train them to have a critical view and raise awareness about
misinformation. In the long run, the impact can only be positive as even without the SAMS-HITL app
people will think twice before passing news along.</p>
      <p>A limitation in the results presented is the size of the dataset and the potential bias in the classes
due to the limited diversity of sources and the length of the text in the news articles. Validating
SAMSHITL will call for its application on a much larger dataset. Having more data gives opportunity to
consider word-embedding techniques for feature extraction and application of deep learning models
for classification. Future work intends to further investigate the SAMS features. One direction is to
explore the potential of automatically answering the questions related to author and spelling, which
could reduce the human efort. Automated tools in combination with a customized text processing
algorithm can be used to identify grammar mistakes and generate a score for the spelling. For news
articles published on web portals, identifying and extracting metadata information related to the
author could be done automatically. However, this is challenging for news stories published and shared
via diferent social network channels. Further work on SAMS features will investigate the impact of
using a score range instead of the binary yes/no values.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work was partly funded by the HES-SO (project no. 104353) and Hasler Foundation in the context
of the project City-Stories (contract no. 17055).
[21] T. De Smedt, W. Daelemans, Pattern for python, The Journal of Machine Learning Research 13
(2012) 2063–2067.
[22] M. Allahbakhsh, B. Benatallah, A. Ignjatovic, H. R. Motahari-Nezhad, E. Bertino, S. Dustdar,</p>
      <p>Quality control in crowdsourcing systems: Issues and directions, Internet Computing (2013).
[23] D. Difallah, E. Filatova, P. Ipeirotis, Demographics and dynamics of mechanical turk workers,
in: Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining,
WSDM ’18, Association for Computing Machinery, New York, NY, USA, 2018, p. 135–143.
[24] G. Kazai, J. Kamps, N. Milic-Frayling, The face of quality in crowdsourcing relevance labels:
Demographics, personality and labeling accuracy, in: Proceedings of the 21st ACM International
Conference on Information and Knowledge Management, CIKM ’12, 2012.
[25] N. Quoc Viet Hung, N. T. Tam, L. N. Tran, K. Aberer, An evaluation of aggregation techniques
in crowdsourcing, in: Web Information Systems Engineering – WISE 2013, 2013, pp. 1–15.
[26] A. P. Dawid, A. M. Skene, Maximum likelihood estimation of observer error-rates using the em
algorithm, Journal of the Royal Statistical Society: Series C (Applied Statistics) 28 (1979) 20–28.
[27] W. Y. Wang, “liar, liar pants on fire”: A new benchmark dataset for fake news detection, in:</p>
      <p>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017.
[28] W. Ferreira, A. Vlachos, Emergent: a novel data-set for stance classification, in: Proceedings
of the 2016 conference of the North American chapter of the association for computational
linguistics: Human language technologies, 2016, pp. 1163–1168.
[29] J. Golbeck, M. Mauriello, B. Auxier, K. H. Bhanushali, C. Bonk, M. A. Bouzaghrane, C. Buntain,
R. Chanduka, P. Cheakalos, J. B. Everett, et al., Fake news vs satire: A dataset and analysis, in:
Proceedings of the 10th ACM Conference on Web Science, 2018, pp. 17–21.
[30] J. Thorne, A. Vlachos, C. Christodoulopoulos, A. Mittal, Fever: a large-scale dataset for fact
extraction and verification, in: Proceedings of the 2018 Conference of the North American
Chapter of the Association for Computational Linguistics, 2018, pp. 809–819.
[31] N. Ruchansky, S. Seo, Y. Liu, Csi: A hybrid deep model for fake news detection, in: Proceedings
of the 2017 ACM on Conference on Information and Knowledge Management, 2017, pp. 797–806.
[32] V. L. Rubin, N. Conroy, Y. Chen, S. Cornwell, Fake news or truth? using satirical cues to
detect potentially misleading news, in: Proceedings of the second workshop on computational
approaches to deception detection, 2016, pp. 7–17.
[33] V. Pérez-Rosas, B. Kleinberg, A. Lefevre, R. Mihalcea, Automatic detection of fake news, in:</p>
      <p>Proceedings of the 27th International Conference on Computational Linguistics, 2018.
[34] S. Tschiatschek, A. Singla, M. Gomez Rodriguez, A. Merchant, A. Krause, Fake news detection in
social networks via crowd signals, in: The Web Conference 2018, WWW ’18, 2018, p. 517–524.
[35] K. Roitero, M. Soprano, B. Portelli, D. Spina, V. Della Mea, G. Serra, S. Mizzaro, G. Demartini, The
covid-19 infodemic: Can the crowd judge recent misinformation objectively?, in: Proceedings
of the 29th ACM International Conference on Information &amp; Knowledge Management, 2020.
[36] Y. Li, T. Grandison, P. Silveyra, A. Douraghy, X. Guan, T. Kieselbach, C. Li, H. Zhang, Jennifer
for COVID-19: An NLP-powered chatbot built for the people and by the people to combat
misinformation, in: Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020, 2020.
[37] S. Ahmed, K. Hinkelmann, F. Corradini, Combining machine learning with knowledge
engineering to detect fake news in social networks-a survey, in: AAAI’19 Spring Symposium, 2019.
[38] S. Shabani, M. Sokhn, Hybrid machine-crowd approach for fake news detection, in: 2018 IEEE
4th International Conference on Collaboration and Internet Computing (CIC), 2018, pp. 299–306.
[39] A. T. Nguyen, A. Kharosekar, M. Lease, B. C. Wallace, An interpretable joint graphical model for
fact-checking from crowds., in: AAAI Conference on Artificial Intelligence, 2018, pp. 1511–1518.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ray</surname>
          </string-name>
          ,
          <article-title>Mapping recent development in scholarship on fake news and misinformation,</article-title>
          <year>2008</year>
          -
          <fpage>2017</fpage>
          :
          <article-title>Disciplinary contribution, topics, and impact, American Behavioral Scientist August (</article-title>
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G. L.</given-names>
            <surname>Ciampaglia</surname>
          </string-name>
          ,
          <article-title>Fighting fake news: a role for computational social science in the fight against digital misinformation</article-title>
          ,
          <source>Journal of Computational Social Science</source>
          <volume>1</volume>
          (
          <year>2018</year>
          )
          <fpage>147</fpage>
          -
          <lpage>153</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zannettou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sirivianos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Blackburn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kourtellis</surname>
          </string-name>
          ,
          <article-title>The web of false information: Rumors, fake news, hoaxes, clickbait , and various other shenanigans</article-title>
          ,
          <source>Journal of Data and Information Quality</source>
          <volume>11</volume>
          (
          <year>2019</year>
          )
          <fpage>37</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Rothkopf</surname>
          </string-name>
          ,
          <article-title>When the buzz bites back</article-title>
          , The Washington Post (
          <year>2003</year>
          )
          <article-title>BO1</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Harman</surname>
          </string-name>
          , Collateral Damage:
          <article-title>The Imperiled Status of Truth in American Public Discourse</article-title>
          and Why it Matters to You, Author House, Bloomington, Indiana,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P. B.</given-names>
            <surname>Brandtzaeg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Følstad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A. C.</given-names>
            <surname>Dominguez</surname>
          </string-name>
          ,
          <article-title>How journalists and social media users perceive online fact-checking and</article-title>
          verification services,
          <source>Journalism Practice</source>
          <volume>12</volume>
          (
          <year>2018</year>
          )
          <fpage>1109</fpage>
          -
          <lpage>1129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Graves</surname>
          </string-name>
          ,
          <source>Understanding the Promise and Limits of Automated Fact-Checking, Report, Reuters Institute for the Study of Journalism</source>
          et University of Oxford,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. B.</given-names>
            <surname>Brandtzaeg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Følstad</surname>
          </string-name>
          ,
          <article-title>Trust and distrust in online fact-checking services</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>60</volume>
          (
          <year>2017</year>
          )
          <fpage>65</fpage>
          -
          <lpage>71</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>[9] Evaluating resources: the crapp test</article-title>
          ,
          <year>2015</year>
          . URL: https://researchguides.ben.edu/c.php?g=
          <volume>261612</volume>
          &amp;p=
          <fpage>2441794</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Konstantinovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Price</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Babakar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zubiaga</surname>
          </string-name>
          , Towards automated factchecking:
          <article-title>Developing an annotation schema and benchmark for consistent automated claim detection</article-title>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kharosekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Krishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Krishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tate</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C.</given-names>
            <surname>Wallace</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lease</surname>
          </string-name>
          ,
          <article-title>Believe it or not: Designing a human-ai partnership for mixed-initiative fact-checking</article-title>
          ,
          <source>in: UIST '18: The 31st Annual ACM Symposium on User Interface Software and Technology</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>189</fpage>
          -
          <lpage>199</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>G.</given-names>
            <surname>Demartini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mizzaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <article-title>Human-in-the-loop artificial intelligence for fighting online misinformation: Challenges and opportunities</article-title>
          ,
          <source>The Bulletin of the Technical Committee on Data Engineering</source>
          <volume>43</volume>
          (
          <year>2020</year>
          )
          <fpage>65</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Rehm</surname>
          </string-name>
          ,
          <article-title>An infrastructure for empowering internet users to handle fake news and other online media phenomena</article-title>
          ,
          <source>in: International Conference of the German Society for Computational Linguistics and Language Technology</source>
          , Springer,
          <year>2017</year>
          , pp.
          <fpage>216</fpage>
          -
          <lpage>231</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Esteves</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Chawla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <article-title>Belittling the source: Trustworthiness indicators to obfuscate fake news on the web</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Fact Extraction and VERification (FEVER)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>59</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ranganathan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Metz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Appling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Sehat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gilmore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. B.</given-names>
            <surname>Adams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Vincent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Robbins</surname>
          </string-name>
          , et al.,
          <article-title>A structured response to misinformation: Defining and annotating credibility indicators in news articles</article-title>
          ,
          <source>in: The Web Conference</source>
          <year>2018</year>
          ,
          <year>2018</year>
          , p.
          <fpage>603</fpage>
          -
          <lpage>612</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>D. M. Lazer</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Baum</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Benkler</surname>
            ,
            <given-names>A. J.</given-names>
          </string-name>
          <string-name>
            <surname>Berinsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Greenhill</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Menczer</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          <string-name>
            <surname>Metzger</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Nyhan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Pennycook</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Rothschild</surname>
          </string-name>
          , et al.,
          <article-title>The science of fake news</article-title>
          ,
          <source>Science</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <article-title>Library guides: News literacy</article-title>
          ,
          <year>Accessed 2020</year>
          . URL: https://nwtc.libguides.com/news.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Spring</surname>
          </string-name>
          ,
          <article-title>Coronavirus: The seven types of people who start and spread viral misinformation</article-title>
          ,
          <source>BBC Trending</source>
          (
          <year>2020</year>
          ). URL: https://www.bbc.com/news/blogs-trending-
          <volume>52474347</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>K.</given-names>
            <surname>Shu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sliva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          , H. Liu,
          <article-title>Fake news detection on social media: A data mining perspective</article-title>
          ,
          <source>SIGKDD Explor. Newsl</source>
          .
          <volume>19</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>L.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          , CoAID: COVID-19 Healthcare Misinformation Dataset, https://arxiv.org/abs/
          <year>2006</year>
          .00885,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>