<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Hybrid Recognition System for Check-worthy Claims Using Heuristics and Supervised Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chaoyuan Zuo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ayla Ida Karakas</string-name>
          <email>ayla.karakas@stonybrook.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ritwik Banerjee</string-name>
          <email>rbanerjeeg@cs.stonybrook.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Linguistics Stony Brook University</institution>
          ,
          <addr-line>Stony Brook New York 11794</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In recent years, the speed at which information disseminates has received an alarming boost from the pervasive usage of social media. To the detriment of political and social stability, this has also made it easier to quickly spread false claims. Due to the sheer volume of information, manual fact-checking seems infeasible, and as a result, computational approaches have been recently explored for automated fact-checking. In spite of the recent advancements in this direction, the critical step of recognizing and prioritizing statements worth fact-checking has received little attention. In this paper, we propose a hybrid approach that combines simple heuristics with supervised machine learning to identify claims made in political debates and speeches, and provide a mechanism to rank them in terms of their \check-worthiness". The viability of our method is demonstrated by evaluations on the English language dataset as part of the Check-worthiness task of the CLEF-2018 Fact Checking Lab.</p>
      </abstract>
      <kwd-group>
        <kwd>Check-worthiness</kwd>
        <kwd>Feature Selection</kwd>
        <kwd>Stylometry</kwd>
        <kwd>Multi-layer Perceptron</kwd>
        <kwd>Heuristics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        It is no secret that we live in an age of ubiquitous web and social media. For
the most part, any Internet user readily acquires the latent power of civilian
commentary and journalism [
        <xref ref-type="bibr" rid="ref10 ref3">3,10</xref>
        ]. Consequently, information available on the
web now carries the potential to propagate amid the public domain with
unprecedented speed and reach. The ordinary Internet user, however, contends with an
overwhelming amount of information, which makes the task of determining the
accuracy and integrity of the claims all the more onerous. Additionally, users
usually want their beliefs to be con rmed by information [
        <xref ref-type="bibr" rid="ref18 ref34">18,34</xref>
        ]. The con uence
of vast amounts of information and such con rmation bias, thus, can create a
society where unveri ed information runs amok masquerading as facts. While
correcting con rmation biases at a social scale may be extremely challenging and
even controversial, the spread of misinformation can be mitigated by focusing
only on curating the claims.
      </p>
      <p>
        Comprehensive manual fact-checking is highly tedious and, in light of the sheer
volume of information, infeasible. To overcome this hurdle, several approaches to
automated fact-checking have been proposed in the nascent eld of computational
journalism [
        <xref ref-type="bibr" rid="ref5 ref8">5,8</xref>
        ]. Some prior work took to computing the semantic similarity
between claims [
        <xref ref-type="bibr" rid="ref13 ref4">4,13</xref>
        ], while others proposed fact-checking as a question-answering
task [
        <xref ref-type="bibr" rid="ref33 ref5">5,33,36</xref>
        ]. Both approaches need to extract statements to be fact-checked
before the actual veri cation process can begin. ClaimBuster [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] was the rst
fact-checking system that assigned to each sentence a check-worthiness score
between 0 and 1. Subsequently, a multi-class classi cation approach with fewer
features was explored to speci cally identify check-worthy claims, but it su ered
from comparatively lower precision [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. Outside of this small body of work, the
preliminary step of identifying check-worthy claims has received little attention.
Gencheva et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] were the rst to develop a publicly available dataset for this
task. Their annotations were obtained from nine fact-checking websites. They also
used a signi cantly richer feature set. Keeping in line with the observations made
by prior work regarding the extent of overlap in lexical and shallow syntactic
features [
        <xref ref-type="bibr" rid="ref20 ref9">9,20</xref>
        ], we use a signi cantly richer set of features derived from word
embeddings and deep syntactic structures.
      </p>
      <p>
        In this work, our focus is on recognizing \check-worthy" statements. Accurate
identi cation of such statements will bene t the fact-checking and veri cation
processes that follow, independent of the speci c techniques used therein. We use
the task formulation, data, and evaluation framework provided by the
CLEF2018 Lab on Automatic Identi cation and Veri cation of Claims in Political
Debates [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] as part of their rst task { Check-Worthiness [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Task, Data, and Evaluation Framework</title>
      <p>The CLEF 2018 Fact Checking Lab designed two tasks that, when put together,
form the complete fact-checking pipeline. In this work, however, we focus
exclusively on the rst.
2.1</p>
      <sec id="sec-2-1">
        <title>The Task: Check-Worthiness</title>
        <p>The rst task { check-worthiness { was de ned by the CLEF 2018 Fact Checking
Lab as follows:</p>
        <p>
          Predict which claim in a political debate should be prioritized for
factchecking. In particular, given a debate, the goal is to produce a ranked
list of its sentences based on their worthiness for fact checking [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
The goal of this task is to automatically recognize claims worth checking, and
present them in order of priority (i.e., as a ranked list of claims) to journalists
or even ordinary Internet and social media users. The ranking is attained in
terms of a check-worthiness score. This approach helps the recipient tackle the
problem of information overload and instead, directly focus on the most important
statements. The output, therefore, can be fed to an automated fact-checker or
be used in a manual pursuit of veri cation. Either way, it can raise awareness of
individual users and stymie the dissemination of false claims in social media.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Data</title>
        <p>Given the alleged impact of disinformation and `fake news' on the 2016 US
presidential election, and the controversy surrounding it, any data pertaining to
this election cycle is extremely relevant in terms of fact-checking endeavors having
a positive social and political impact in the future. As such, a political debate
dataset was provided in English and Arabic. Since our methodology involves
heuristics that rely on linguistic insight, we used the English language dataset.</p>
        <p>The training data comprised three political debates. Each debate was split
into sentences, and each sentence was associated with its speaker and annotated
by experts as check-worthy or not (labeled 1 and 0, respectively). This data
contained a total of 3,989 sentences, of which only 94 were labeled as check-worthy
{ a staggering imbalance with only 2.36% of the dataset bearing the label of the
target class. A few simple sentences from this training data, along with their
speakers and labels, are presented in Table 1.</p>
        <p>The test data was a collection of two political debates and ve political
speeches.3 The total number of sentences in these two categories (Debate and
Speech) were 2,815 and 2,064, respectively.</p>
        <p>In this work, we did not employ any external knowledge other than
domainindependent language resources such as parsers and lexicons. Instead, we focused
extracting linguistic features indicative of check-worthiness.
3 The lab task provided all seven les together, without this categorization into speeches
and debates. We, however, chose to treat these di erently since language use is very
di erent in these two scenarios: debates consist of the interactive statements made
by the candidates and the moderator, while speeches only have a single speaker, and
there is no two-sided conversational structure.
2.3</p>
      </sec>
      <sec id="sec-2-3">
        <title>Evaluation Framework</title>
        <p>The evaluation was done on the test data provided as part of the task. This data
was released much later to the participants, with the gold standard labels for the
sentences in the test data withheld. Once we selected the models, we ran it on
the entire test data, and used average precision to measure the quality of the
output ranking. Average precision is de ned as</p>
        <p>AP =
1 n</p>
        <p>
          X Prec(k) (k)
nchk k=1
where nchk is the number of check-worthy sentences, n is the total number of
sentences, Prec(k) is the precision at cut-o k in the list of sentences ranked by
check-worthiness, and (k) is the indicator function equaling 1 if the sentence
at rank k is check-worthy, and 0 otherwise. The primary metric used by the
Fact Checking Lab [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] for the check-worthiness task was mean average precision
(MAP), de ned simply as the mean of the average precisions over all queries.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>Our methodology is a hybrid of rule-based heuristics and supervised classi cation.
The motivation for this approach was to test the extent to which check-worthiness
can be determined based on language constructs without relying on encyclopedic
knowledge. Moreover, our aim was to develop an approach that was not speci c
to the domain of politics. In this section, we describe the data processing, feature
selection, and heuristics involved in building our classi cation models.
3.1</p>
      <sec id="sec-3-1">
        <title>Data Processing</title>
        <p>The rst step of our processing involved normalizing the speaker names. We did
this by adding speaker-speci c rules in order to correctly match the speakers
extracted from various sentences to the actual speakers associated with the
sentences. For example, speakers in the test data included \Hillary Clinton
(D-NY)", \Former Secretary of State, Presidential Candidate", and
simply \Clinton". These are, of course, all referring to the same speaker.</p>
        <p>Next, we noted that the training data consisted only of political debates
where multiple entities (two political candidates, a moderator, and the occasional
audience reaction) engage in a conversation. Due to the very nature of debates,
the rhetorical structure is di erent from speeches delivered by a single speaker.
The test data, however, also included political speeches. Therefore, we extracted
all sentences attributed to a speaker to create sub-datasets. This formed a new
training sample, which we then used to train models to identify check-worthy
sentences from speeches4. To identify check-worthy sentences from political
debates, we used the original training data to train the models.
4 The provided training sample included two speeches, and both were by Donald
Trump. As a result, for the purpose of this task, a single sub-dataset was created.
The approach is independent of the speaker and the number of speakers, however.
For both speeches and debates, we extracted a set of syntactic and semantic
features to obtain a consistent knowledge representation, and converted every
sentence into a vector in an abstract semantic space. The details of these features
and the resultant feature vector are discussed below.</p>
        <p>
          Sentence Embedding: Traditional supervised learning in natural language
processing tasks have used vector spaces where dimensions correspond to words
(or other linguistic units). This, however, is not in accordance with the well-known
distributional hypothesis in linguistics: words that occur in similar contexts tend
to have similar meanings [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. This necessitates the representation of sentences
in a low-dimensional semantic space where similar meanings are closer together.
Modeling sentence meanings in a low-dimensional space is a topic of extensive
research by itself, and beyond the scope of this work. Instead, we adopted a
simple method that leverages word embeddings. We used the 300-dimensional
pretrained Google News word embeddings5 to represent each word as a vector [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ],
and took the arithmetic mean of all such vectors corresponding to the words in a
sentence to obtain an abstract sentence embedding.
        </p>
        <p>
          Lexical Features: From the training data, we removed stopwords and stemmed
the remaining terms using the Snowball stemmer [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ].
        </p>
        <p>
          Stylometric Features: Stylometry, the statistical analysis of variations in
linguistic constructs, has been used with great success in distinguishing deceptive
from truthful language [
          <xref ref-type="bibr" rid="ref26 ref6">6,26</xref>
          ], and objective from subjective remarks [
          <xref ref-type="bibr" rid="ref19 ref21">19,21</xref>
          ].
Accordingly, we surmised that capturing stylistic variation will aid in the
identi cation of check-worthy sentences as well, especially since they are typically
expected to appear factual and objective.
        </p>
        <p>In order to obtain shallow syntactic features from each sentence, we extracted
the part-of-speech (POS) tags, the total number of tokens, and the number of
tokens in past, present, and future tenses. We were able to infer the tense from
the POS tags (e.g., both vbd and vbz are verb tags, but they indicate past
and present tense, respectively). Additionally, we also extracted the number of
negations in each sentence.
5 Available at https://code.google.com/archive/p/word2vec/.</p>
        <p>
          More complex structural patterns of language, however, can only be captured
by deep syntactic features. For that, we generated the constituency parse trees
of all sentences, and selected clause-level and phrase-level tags. The number of
words within the scope of each tag were included as the corresponding feature
values. These tags, as de ned in the Penn Treebank [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], are shown in Table 2. In
addition to stylometry, the motivation behind using the number of words was
to obtain a representation of the amount of information available under speci c
syntactic structures. Fig. 1 illustrates this point with the parse tree of a sentence
from the training data that was labeled as check-worthy.
        </p>
        <p>
          Semantic Features: We used the Stanford named entity recognizer (NER) [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]
to extract the number of named entities in a sentence. Additionally, we appended
an extra feature for named entities of the type person.
        </p>
        <p>
          A ective Features: We used the TextBlob [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] library to train a nave Bayes
classi er on the pioneering movie review corpus for sentiment analysis [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ],
and thereby obtained a sentiment score for each sentence. In addition to overt
sentiment, we also used the connotation of words in a sentence as features. For this,
we employed Connotation WordNet [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], which assigns a (positive or negative)
connotation score to each word. For every sentence, we queried this lexicon and
retrieved the connotation score of its words. Finally, the overall connotation of
the sentence was attributed simply to the mean of these scores.
        </p>
        <p>
          Additionally, we also utilized lexicons that contain information about the
subjective or objective nature of words [35], whether they directly indicate or are
typically associated with language that indicates bias [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ], and whether they are
typically used to voice positive or negative opinions [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. For every sentence, we
extracted the number of words in these categories (as de ned by their scores in
these lexicons), thus forming four new features: (i) subjectivity, (ii) direct bias,
(iii) associated bias, and (iv) opinion.
        </p>
        <p>
          Metadata Features: In addition to the syntactic and semantic features
described above, we also included three binary non-linguistic features extracted
from the training sample, indicating whether or not (i) the speaker's opponent
is mentioned, (ii) the speaker is the anchor/moderator, or (iii) the sentence is
immediately followed by intense reaction. The third feature is encoded in the
training data as a `system' reaction, as shown by the last sentence in Table 1.
Discourse Features: All the above features were extracted without regards
to the category (i.e., Debate and Speech). Since debates involve an interactive
discourse structure where sentences are often formed as an immediate response to
statements made by others, we include segments from the debates. We adopt the
approach taken by Gencheva et al [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] and regard a \segment" to be the maximal
set of consecutive sentences by the same speaker. As features, we include the
relative position of a sentence within its segment, and the number of sentences
in the previous, current and subsequent segments.
        </p>
        <p>
          Feature Selection The feature extraction processes described above yielded a
very high-dimensional feature space. High-dimensional spaces, however, quickly
lead to a decrease in the predictive power of models [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ]. Moreover, given the
extreme class imbalance, classi cation in such a space is likely to ignore important
features indicative of the minority class (in this case, the `check-worthy' sentences).
        </p>
        <p>
          To reduce the dimensionality, we applied a feature selection module using
the scikit-learn library [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]. As the rst step, univariate feature selection was
performed, and the 2,000 best features were selected based on 2-test. Next,
armed with the observation that linear predictive models with L1 loss yield
sparse solutions and encourage the vanishing coe cients for weakly correlated
features [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], we used a support vector machine (SVM) model with linear kernel
and L1 regularization to further remove the relatively unimportant features. This
step was rst done on the entire training data, and then combined with repeated
undersampling (without replacement) for the majority class. Each iteration of
this undersampling process resulted in a small but balanced training sample.
A L1-regularized SVM learner was trained on every sample generated in this
manner, and features with vanishing coe cients were discarded. The cumulative
e ect of these feature selection steps was a reduction of the feature space to 2,655
and 2,404 dimensions for identi cation of check-worthy claims from debates and
speeches, respectively.
3.3
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Heuristics</title>
        <p>Certain heuristics were introduced to override the scores assigned by the
classi cation models. These rules di ered slightly based on (i) the category, i.e.,
speech or debate, and (ii) whether or not the `strict' heuristics were deployed.
Algorithm 1 Heuristics for assigning the check-worthiness score w( ) to sentences.
Require: category 2 fspeech; debateg,
strict mode 2 ftrue, falseg, sentence S.
min token count 0
if category is speech then
if strict mode then</p>
        <p>min token count
else</p>
        <p>min token count
end if
else
if strict mode then</p>
        <p>min token count
else</p>
        <p>min token count
end if
end if
10
8
7
5
if Sspeaker is system then</p>
        <p>w(s) 10 8
end if
if Snumber of tokens &lt; min token count
then</p>
        <p>w(s) 10 8
end if
if S contains \thank you" then</p>
        <p>w(s) 10 8
end if
if Snumber of subjects &lt; 1 then
if category is speech then</p>
        <p>w(s) 10 8
else if S contains \?" then</p>
        <p>w(s) 10 8
end if
end if
The strictness ag was introduced to control the threshold sentence size. When
active, it would tend to discard more sentences.</p>
        <p>These rules are speci ed in Algorithm 1. One particular rule required the
identi cation of subjects in a sentence. To extract this information, we generated
dependency parse trees of the sentences and counted the number of times any of
the following dependency labels appeared: nsubj, csubj, nsubjpass, csubjpass,
or xsubj. The rst two indicate nominal and clausal subjects, respectively. The
next two indicate nominal and clausal subjects in a passive clause, and the last
label denotes a controlling subject, which relates an open clausal complement to
its external clause.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Models</title>
      <p>Our experiments comprised two supervised learning algorithms: support vector
machines (SVM) and multilayer perceptrons (MLP). Additionally, we also built
an ensemble model combing the two. In this section, we provide a description of
these three models, along with their training processes.</p>
      <p>
        For reasons described in Sec. 3.2, the SVM utilized a linear kernel with L1
regularization for feature selection. However, due to the propensity of the L1 loss
function to miss optimal solutions, we used L2 loss in building the nal model
after completing feature selection. Our second model was the MLP. Here, we
used two hidden layers with 100 units and 8 units in them, respectively. We used
the hyperbolic tangent (tanh) as our activation function since it achieved better
results when compared to recti ed linear units (ReLU). Stochastic optimization
was done with Adam [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. To avoid over tting, we used L2-regularization in both
MLP? 0.1332
MLPstr 0.1366
ENS 0.1317
MLPnone 0.1086
SVM and MLP. Third, we built an ensemble model that combines SVM and
MLP (without the strict heuristics). In this model, the nal output score was a
normalization (by standard deviation) of the results of SVM and MLP, and then
computing the average.
      </p>
      <p>
        For all three models, class imbalance was a hindrance during the training
process. To overcome that, we used ADASYN[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], an adaptive synthetic sampling
algorithm for imbalanced learning. For model selection, we used 3-fold
crossvalidation for debates, using two les for training and the remaining one for
testing, to evaluate model performances and tune parameters. For speeches,
we split the training sample into two halves (one le in each) for 2-fold
crossvalidation. The evaluation script was provided by the task organizers, with the
mean average precision (MAP) being the primary evaluation metric.
      </p>
      <p>MLP without the strict heuristics demonstrated the best results during the
training process, so this was submitted for the primary run. For the two contrastive
runs, we submitted (i) MLP with strict heuristics, and (ii) the ensemble model
without the strict heuristics.</p>
    </sec>
    <sec id="sec-5">
      <title>Results and Analysis</title>
      <sec id="sec-5-1">
        <title>Empirical Results</title>
        <p>The detailed performance of all three submissions we made is shown in Table 3.
Even though MLP yielded the best training results without the strict heuristics,
MLPstr performed demonstrably better across multiple metrics on the test data.
Our third model, the ensemble classi er, performed poorly in general compared
to both MLP models. It did, however, achieve slightly better mean R-precision
and mean precision at higher cuto s (k = 10 and 50).</p>
        <p>Without the inclusion of any heuristics, the performance of MLP dropped
signi cantly. This was expected, since the heuristics were designed to address the
aws of the classi ers. This model was not among the submissions, but we include
it here for comparison. The di erence between MLP and MLPnone quanti es the
extent to which the rules help the supervised learners.</p>
        <p>Next, in Table 4, we present the comparison between the results obtained
by all participants. This comparison was done only on the primary submission
from each team. Our MLP model without the strict heuristics achieved the best
MAP, MRR, and MRP scores. Further, it also outperformed the others in terms
of correctly placing the check-worthy sentences at the very top of the ranked
output list, as demonstrated by the mean precision at low values (k = 1 and 3).
Identifying check-worthy sentences is a di cult and novel task, and even the best
model su ered from misclassi cation errors. Upon analyzing such mistakes made
by the MLP models, we were able to discern a few reasons.</p>
        <p>First, tense plays a logical role in check-worthiness, since future actions cannot
be veri ed. However, the part-of-speech tagging often confuses the future tense
with the present continuous (e.g., \We're cutting taxes."). Second, we observed
that anecdotal stories are often highly prioritized as check-worthy, while they are
not. These sentences are usually complex, with a lot of content, which makes it
easy for the model to con ate them with other complex sentences pertaining to
real events deemed check-worthy. Third, the presence of duplicate sentences in
the data means that a misclassi cation gets ampli ed, while the presence of very
similar sentences with di erent labels likely makes the feature selection stage
discard potentially useful features.</p>
        <p>At a more abstract level, rhetorical gures of speech play a critical role.
They often break the structures associated with standard sentence formation.
Several sentences that were misclassi ed exhibited constructs such as scesis
onomaton, where words or phrases with nearly equivalent meaning are repeated.
We conjecture that this makes the model falsely believe that there is more
informational content in the sentence. Such gures of speech become even harder
to handle when they occur across multiple speakers in debates. The conversational
aspect of debates also causes another problem: quite a few sentences are short,
and in isolation, would perhaps not be check-worthy. However, as a response to
things mentioned earlier in the debate, they are.</p>
        <p>Another complex issue leading to misclassi cation is the use of sentence
fragments. This is sparingly used for dramatic e ect in literature, but was seen
with alarming frequency in the political debates due to the prevalence of
illformed or partly-formed sentences stopping and then giving way to another
sentence. In some cases, the fragments are portions of the sentence that the
speaker repeats. An example of such a fragment is the sentence \Ambassador
Stevens { Ambassador Stevens sent 600 requests for help.", where the phrase
\Ambassador Stevens" is repeated.</p>
        <p>A proper approach to deal with these hurdles is a complex matter in and by
itself. We believe that our features are better suited for written language than
speech or debate transcripts. In the presence of signi cantly more labeled data
for check-worthiness, ablation studies that remove such sentences could provide
empirical evidence of this intuition.
6</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion and Future Work</title>
      <p>We developed a hybrid system that combines a few rules with supervised learning
to detect check-worthy sentences in political debates and speeches. To tackle the
severity of class imbalance, our development also included a sophisticated feature
selection process and special sampling methods. Our primary model achieved the
best results among all participants over multiple performance metrics.</p>
      <p>This work opens up several intriguing possibilities for future research in the
eld of fact-checking. First, we intend to study in greater details the linguistic
forms of informational content. Shallow syntax has been explored to understand
this aspect of language in sociolinguistics, and some work has even looked into deep
syntactic features. This approach has, however, not yet been applied to identifying
check-worthy sentences. Furthermore, more complex neural network structures
need to be thoroughly investigated. Along this line, we will be investigating deep
learning models with feedback control. A stringent and focused work on these
issues will empower journalists and citizens alike to be better informed and more
cognizant of false claims permeating news and social media now. To that end,
we also need complementary advances in related areas like natural language
querying, crowdsourcing, source identi cation, and social network analysis.
Acknowledgment: This work was supported in part by the U.S. National
Science Foundation (NSF) under the award SES-1834597.
35. Wilson, T., Wiebe, J., Ho mann, P.: Recognizing Contextual Polarity in
PhraseLevel Sentiment Analysis. In: Proceedings of the Conference on Human Language
Technology and Empirical Methods in Natural Language Processing. pp. 347{354.</p>
      <p>Association for Computational Linguistics (2005)
36. Wu, Y., Agarwal, P.K., Li, C., Yang, J., Yu, C.: Toward computational fact-checking.</p>
      <p>Proceedings of the VLDB Endowment 7(7), 589{600 (2014)</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Atanasova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marquez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <article-title>Barron-Ceden~o,</article-title>
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Kyuchukov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            , Da San Martino, G.,
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>Overview of the CLEF-</article-title>
          2018
          <source>CheckThat! Lab on Automatic Identi cation and Veri cation of Political Claims, Task</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Check-Worthiness</surname>
            . In: Cappellato,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>J.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Soulier</surname>
            ,
            <given-names>L</given-names>
          </string-name>
          . (eds.)
          <source>CLEF 2018 Working Notes. Working Notes of CLEF 2018 - Conference and Labs of the Evaluation Forum. CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Avignon, France (
          <year>September 2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bies</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferguson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>MacIntyre</given-names>
            , R.,
            <surname>Tredinnick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Marcinkiewicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.A.</given-names>
            ,
            <surname>Schasberger</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          :
          <article-title>Bracketing Guidelines for Treebank II Style Penn Treebank Project</article-title>
          . University of Pennsylvania 97,
          <issue>100</issue>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bruns</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , High eld, T.:
          <article-title>Blogs, Twitter, and breaking news: the produsage of citizen journalism</article-title>
          .
          <source>In: Produsing Theory in a Digital World: The Intersection of Audiences and Production in Contemporary Theory</source>
          , vol.
          <volume>80</volume>
          , pp.
          <volume>15</volume>
          {
          <fpage>32</fpage>
          .
          <string-name>
            <surname>Peter Lang Publishing Inc.</surname>
          </string-name>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cazalens</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lamarre</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leblay</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manolescu</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tannier</surname>
            ,
            <given-names>X.:</given-names>
          </string-name>
          <article-title>A content management perspective on fact-checking</article-title>
          . In:
          <article-title>" Journalism, Misinformation and Fact Checking" alternate paper track of"</article-title>
          <source>The Web Conference"</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Cohen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Computational Journalism: A Call to Arms to Database Researchers</article-title>
          .
          <source>In: Conference on Innovative Data Systems Research. CIDR '11</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Asilomar, California, USA (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Syntactic Stylometry for Deception Detection</article-title>
          .
          <source>In: Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers-Volume</source>
          <volume>2</volume>
          . pp.
          <volume>171</volume>
          {
          <fpage>175</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Finkel</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grenager</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Incorporating non-local information into information extraction systems by Gibbs sampling</article-title>
          .
          <source>In: Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics</source>
          . pp.
          <volume>363</volume>
          {
          <fpage>370</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Flew</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spurgeon</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daniel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Swift</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>The promise of computational journalism</article-title>
          .
          <source>Journalism Practice</source>
          <volume>6</volume>
          (
          <issue>2</issue>
          ),
          <volume>157</volume>
          {
          <fpage>171</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Gencheva</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marquez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <article-title>Barron-Ceden~o,</article-title>
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Koychev</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.:</surname>
          </string-name>
          <article-title>A contextaware approach for detecting worth-checking claims in political debates</article-title>
          .
          <source>In: Proceedings of the International Conference Recent Advances in Natural Language Processing</source>
          ,
          <string-name>
            <surname>RANLP</surname>
          </string-name>
          <year>2017</year>
          . pp.
          <volume>267</volume>
          {
          <issue>276</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Goode</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Social news, citizen journalism and democracy</article-title>
          .
          <source>New media &amp; society 11(8)</source>
          ,
          <volume>1287</volume>
          {
          <fpage>1305</fpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Harris</surname>
            ,
            <given-names>Z.S.: Distributional</given-names>
          </string-name>
          <string-name>
            <surname>Structure</surname>
          </string-name>
          .
          <source>Word</source>
          <volume>10</volume>
          (
          <issue>2-3</issue>
          ),
          <volume>146</volume>
          {
          <fpage>162</fpage>
          (
          <year>1954</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tremayne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Detecting check-worthy factual claims in presidential debates</article-title>
          .
          <source>In: Proceedings of the 24th ACM International Conference on Information and Knowledge Management</source>
          . pp.
          <year>1835</year>
          {
          <year>1838</year>
          . CIKM (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , G.,
          <string-name>
            <surname>Arslan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caraballo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jimenez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gawsane</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joseph</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kulkarni</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nayak</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          , et al.:
          <article-title>ClaimBuster: The First-ever End-to-end Fact-checking System</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>10</volume>
          (
          <issue>12</issue>
          ),
          <year>1945</year>
          {
          <year>1948</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>He</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bai</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>E.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>ADASYN: Adaptive Synthetic Sampling Approach for Imbalanced Learning</article-title>
          .
          <source>In: Proceedings of the IEEE Joint Conference on Neural Networks (IJCNN)</source>
          ,
          <year>2008</year>
          . pp.
          <volume>1322</volume>
          {
          <fpage>1328</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Hu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Mining and Summarizing Customer Reviews</article-title>
          .
          <source>In: Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>168</volume>
          {
          <fpage>177</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akoglu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
          </string-name>
          , Y.:
          <article-title>ConnotationWordNet: Learning Connotation over the Word+Sense Network</article-title>
          .
          <source>In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics</source>
          (Vol.
          <article-title>1: Long Papers)</article-title>
          . pp.
          <volume>1544</volume>
          {
          <fpage>1554</fpage>
          . Association for Computational Linguistics (
          <year>June 2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ba</surname>
          </string-name>
          , J.:
          <article-title>Adam: A method for stochastic optimization</article-title>
          .
          <source>arXiv preprint arXiv:1412.6980</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Klayman</surname>
          </string-name>
          , J.:
          <article-title>Varieties of Con rmation Bias</article-title>
          . In:
          <article-title>Psychology of learning and motivation</article-title>
          , vol.
          <volume>32</volume>
          , pp.
          <volume>385</volume>
          {
          <fpage>418</fpage>
          .
          <string-name>
            <surname>Elsevier</surname>
          </string-name>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Lamb</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paul</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dredze</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Separating Fact from Fear: Tracking Flu Infections on Twitter</article-title>
          .
          <source>In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>789</volume>
          {
          <issue>795</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>D.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vu</surname>
            ,
            <given-names>N.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blessing</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Towards a text analysis system for political debates</article-title>
          .
          <source>In: Proceedings of the 10th SIGHUM Workshop on Language Technology for Cultural Heritage</source>
          ,
          <source>Social Sciences, and Humanities</source>
          . pp.
          <volume>134</volume>
          {
          <issue>139</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Lex</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ju</surname>
            <given-names>nger</given-names>
          </string-name>
          , A.,
          <string-name>
            <surname>Granitzer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Objectivity Classi cation in Online Media</article-title>
          .
          <source>In: Proceedings of the 21st ACM Conference on Hypertext and Hypermedia</source>
          . pp.
          <volume>293</volume>
          {
          <fpage>294</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Loria</surname>
          </string-name>
          , S.: TextBlob: Simpli ed Text Processing. http://textblob.readthedocs.org/ en/dev/ (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>arXiv preprint arXiv:1301.3781</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Barron-Ceden~o,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Marquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Gencheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Kyuchukov</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          , Da San Martino, G.:
          <article-title>Overview of the CLEF-2018 Lab on Automatic Identi cation and Veri cation of Claims in Political Debates</article-title>
          . In: Working Notes of CLEF 2018 {
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          . CLEF '
          <volume>18</volume>
          ,
          <string-name>
            <surname>Avignon</surname>
          </string-name>
          , France (
          <year>September 2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Ng</surname>
          </string-name>
          , A.Y.:
          <article-title>Feature selection, l 1 vs. l 2 regularization, and rotational invariance</article-title>
          .
          <source>In: Proceedings of the twenty- rst international conference on Machine learning</source>
          . p.
          <fpage>78</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cardie</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hancock</surname>
            ,
            <given-names>J.T.</given-names>
          </string-name>
          :
          <article-title>Finding deceptive opinion spam by any stretch of the imagination</article-title>
          .
          <source>In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies</source>
          { Vol.
          <volume>1</volume>
          . pp.
          <volume>309</volume>
          {
          <fpage>319</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Pang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vaithyanathan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Thumbs up?: sentiment classi cation using machine learning techniques</article-title>
          .
          <source>In: Proceedings of the ACL-02 conference on Empirical methods in natural language processing-</source>
          Volume
          <volume>10</volume>
          . pp.
          <volume>79</volume>
          {
          <fpage>86</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Patwari</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldwasser</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bagchi</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>TATHYA: A Multi-Classi er System for Detecting Check-Worthy Statements in Political Debates</article-title>
          .
          <source>In: Proceedings of the 26th ACM International Conference on Information and Knowledge Management</source>
          . pp.
          <volume>1</volume>
          {
          <issue>4</issue>
          .
          <string-name>
            <surname>CIKM</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Porter</surname>
            ,
            <given-names>M.F.</given-names>
          </string-name>
          :
          <article-title>Snowball: A Language for Stemming Algorithms</article-title>
          . http://snowball. tartarus.org/texts/introduction.html (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Recasens</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Danescu-Niculescu-Mizil</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Linguistic Models for Analyzing and Detecting Biased Language</article-title>
          . In:
          <article-title>Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          .
          <source>vol. 1</source>
          , pp.
          <volume>1650</volume>
          {
          <issue>1659</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Trunk</surname>
            ,
            <given-names>G.V.</given-names>
          </string-name>
          :
          <article-title>A Problem of Dimensionality: A Simple Example</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>1</volume>
          (
          <issue>3</issue>
          ),
          <volume>306</volume>
          {
          <fpage>307</fpage>
          (
          <year>1979</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Vlachos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riedel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Fact Checking:
          <article-title>Task de nition and dataset construction</article-title>
          .
          <source>In: Proceedings of the ACL 2014 Workshop on Language Technologies and Computational Social Science</source>
          . pp.
          <volume>18</volume>
          {
          <issue>22</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Werner</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tankard</surname>
            <given-names>Jr</given-names>
          </string-name>
          ,
          <string-name>
            <surname>J.W.</surname>
          </string-name>
          :
          <article-title>Communication theories: Origins, methods and uses in the mass media (</article-title>
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>