<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Hybrid Recognition System for Check-worthy Claims Using Heuristics and Supervised Learning</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Chaoyuan</forename><surname>Zuo</surname></persName>
							<email>chzuo@cs.stonybrook.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ayla</forename><forename type="middle">Ida</forename><surname>Karakas</surname></persName>
							<email>ayla.karakas@stonybrook.edu</email>
							<affiliation key="aff1">
								<orgName type="department">Department of Linguistics</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ritwik</forename><surname>Banerjee</surname></persName>
							<email>rbanerjee@cs.stonybrook.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="institution">Stony Brook University</orgName>
								<address>
									<postCode>11794</postCode>
									<settlement>Stony Brook New</settlement>
									<region>York</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Hybrid Recognition System for Check-worthy Claims Using Heuristics and Supervised Learning</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B9010D894C2760DD4874DF6E1B15E88E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T02:35+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Check-worthiness</term>
					<term>Multi-layer Perceptron</term>
					<term>Heuristics</term>
					<term>Feature Selection</term>
					<term>Stylometry</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In recent years, the speed at which information disseminates has received an alarming boost from the pervasive usage of social media. To the detriment of political and social stability, this has also made it easier to quickly spread false claims. Due to the sheer volume of information, manual fact-checking seems infeasible, and as a result, computational approaches have been recently explored for automated fact-checking. In spite of the recent advancements in this direction, the critical step of recognizing and prioritizing statements worth fact-checking has received little attention. In this paper, we propose a hybrid approach that combines simple heuristics with supervised machine learning to identify claims made in political debates and speeches, and provide a mechanism to rank them in terms of their "check-worthiness". The viability of our method is demonstrated by evaluations on the English language dataset as part of the Check-worthiness task of the CLEF-2018 Fact Checking Lab.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>It is no secret that we live in an age of ubiquitous web and social media. For the most part, any Internet user readily acquires the latent power of civilian commentary and journalism <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b9">10]</ref>. Consequently, information available on the web now carries the potential to propagate amid the public domain with unprecedented speed and reach. The ordinary Internet user, however, contends with an overwhelming amount of information, which makes the task of determining the accuracy and integrity of the claims all the more onerous. Additionally, users usually want their beliefs to be confirmed by information <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b33">34]</ref>. The confluence of vast amounts of information and such confirmation bias, thus, can create a society where unverified information runs amok masquerading as facts. While correcting confirmation biases at a social scale may be extremely challenging and even controversial, the spread of misinformation can be mitigated by focusing only on curating the claims.</p><p>Comprehensive manual fact-checking is highly tedious and, in light of the sheer volume of information, infeasible. To overcome this hurdle, several approaches to automated fact-checking have been proposed in the nascent field of computational journalism <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b7">8]</ref>. Some prior work took to computing the semantic similarity between claims <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b12">13]</ref>, while others proposed fact-checking as a question-answering task <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b32">33,</ref><ref type="bibr" target="#b35">36]</ref>. Both approaches need to extract statements to be fact-checked before the actual verification process can begin. ClaimBuster <ref type="bibr" target="#b11">[12]</ref> was the first fact-checking system that assigned to each sentence a check-worthiness score between 0 and 1. Subsequently, a multi-class classification approach with fewer features was explored to specifically identify check-worthy claims, but it suffered from comparatively lower precision <ref type="bibr" target="#b27">[28]</ref>. Outside of this small body of work, the preliminary step of identifying check-worthy claims has received little attention. Gencheva et al. <ref type="bibr" target="#b8">[9]</ref> were the first to develop a publicly available dataset for this task. Their annotations were obtained from nine fact-checking websites. They also used a significantly richer feature set. Keeping in line with the observations made by prior work regarding the extent of overlap in lexical and shallow syntactic features <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b19">20]</ref>, we use a significantly richer set of features derived from word embeddings and deep syntactic structures.</p><p>In this work, our focus is on recognizing "check-worthy" statements. Accurate identification of such statements will benefit the fact-checking and verification processes that follow, independent of the specific techniques used therein. We use the task formulation, data, and evaluation framework provided by the CLEF-2018 Lab on Automatic Identification and Verification of Claims in Political Debates <ref type="bibr" target="#b23">[24]</ref> as part of their first task -Check-Worthiness <ref type="bibr" target="#b0">[1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task, Data, and Evaluation Framework</head><p>The CLEF 2018 Fact Checking Lab designed two tasks that, when put together, form the complete fact-checking pipeline. In this work, however, we focus exclusively on the first.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">The Task: Check-Worthiness</head><p>The first task -check-worthiness -was defined by the CLEF 2018 Fact Checking Lab as follows:</p><p>Predict which claim in a political debate should be prioritized for factchecking. In particular, given a debate, the goal is to produce a ranked list of its sentences based on their worthiness for fact checking <ref type="bibr" target="#b8">[9]</ref>.</p><p>The goal of this task is to automatically recognize claims worth checking, and present them in order of priority (i.e., as a ranked list of claims) to journalists or even ordinary Internet and social media users. The ranking is attained in terms of a check-worthiness score. This approach helps the recipient tackle the problem of information overload and instead, directly focus on the most important statements. The output, therefore, can be fed to an automated fact-checker or be used in a manual pursuit of verification. Either way, it can raise awareness of individual users and stymie the dissemination of false claims in social media.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Data</head><p>Given the alleged impact of disinformation and 'fake news' on the 2016 US presidential election, and the controversy surrounding it, any data pertaining to this election cycle is extremely relevant in terms of fact-checking endeavors having a positive social and political impact in the future. As such, a political debate dataset was provided in English and Arabic. Since our methodology involves heuristics that rely on linguistic insight, we used the English language dataset. The training data comprised three political debates. Each debate was split into sentences, and each sentence was associated with its speaker and annotated by experts as check-worthy or not (labeled 1 and 0, respectively). This data contained a total of 3,989 sentences, of which only 94 were labeled as check-worthy -a staggering imbalance with only 2.36% of the dataset bearing the label of the target class. A few simple sentences from this training data, along with their speakers and labels, are presented in Table <ref type="table" target="#tab_0">1</ref>.</p><p>The test data was a collection of two political debates and five political speeches. <ref type="foot" target="#foot_0">3</ref> The total number of sentences in these two categories (Debate and Speech) were 2,815 and 2,064, respectively.</p><p>In this work, we did not employ any external knowledge other than domainindependent language resources such as parsers and lexicons. Instead, we focused extracting linguistic features indicative of check-worthiness.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Evaluation Framework</head><p>The evaluation was done on the test data provided as part of the task. This data was released much later to the participants, with the gold standard labels for the sentences in the test data withheld. Once we selected the models, we ran it on the entire test data, and used average precision to measure the quality of the output ranking. Average precision is defined as</p><formula xml:id="formula_0">AP = 1 n chk n k=1 Prec(k) • δ(k)</formula><p>where n chk is the number of check-worthy sentences, n is the total number of sentences, Prec(k) is the precision at cut-off k in the list of sentences ranked by check-worthiness, and δ(k) is the indicator function equaling 1 if the sentence at rank k is check-worthy, and 0 otherwise. The primary metric used by the Fact Checking Lab <ref type="bibr" target="#b23">[24]</ref> for the check-worthiness task was mean average precision (MAP), defined simply as the mean of the average precisions over all queries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Methodology</head><p>Our methodology is a hybrid of rule-based heuristics and supervised classification. The motivation for this approach was to test the extent to which check-worthiness can be determined based on language constructs without relying on encyclopedic knowledge. Moreover, our aim was to develop an approach that was not specific to the domain of politics. In this section, we describe the data processing, feature selection, and heuristics involved in building our classification models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Data Processing</head><p>The first step of our processing involved normalizing the speaker names. We did this by adding speaker-specific rules in order to correctly match the speakers extracted from various sentences to the actual speakers associated with the sentences. For example, speakers in the test data included "Hillary Clinton (D-NY)", "Former Secretary of State, Presidential Candidate", and simply "Clinton". These are, of course, all referring to the same speaker.</p><p>Next, we noted that the training data consisted only of political debates where multiple entities (two political candidates, a moderator, and the occasional audience reaction) engage in a conversation. Due to the very nature of debates, the rhetorical structure is different from speeches delivered by a single speaker. The test data, however, also included political speeches. Therefore, we extracted all sentences attributed to a speaker to create sub-datasets. This formed a new training sample, which we then used to train models to identify check-worthy sentences from speeches<ref type="foot" target="#foot_1">4</ref> . To identify check-worthy sentences from political debates, we used the original training data to train the models. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Feature Design and Selection</head><p>For both speeches and debates, we extracted a set of syntactic and semantic features to obtain a consistent knowledge representation, and converted every sentence into a vector in an abstract semantic space. The details of these features and the resultant feature vector are discussed below.</p><p>Sentence Embedding: Traditional supervised learning in natural language processing tasks have used vector spaces where dimensions correspond to words (or other linguistic units). This, however, is not in accordance with the well-known distributional hypothesis in linguistics: words that occur in similar contexts tend to have similar meanings <ref type="bibr" target="#b10">[11]</ref>. This necessitates the representation of sentences in a low-dimensional semantic space where similar meanings are closer together.</p><p>Modeling sentence meanings in a low-dimensional space is a topic of extensive research by itself, and beyond the scope of this work. Instead, we adopted a simple method that leverages word embeddings. We used the 300-dimensional pretrained Google News word embeddings<ref type="foot" target="#foot_2">5</ref> to represent each word as a vector <ref type="bibr" target="#b22">[23]</ref>, and took the arithmetic mean of all such vectors corresponding to the words in a sentence to obtain an abstract sentence embedding.</p><p>Lexical Features: From the training data, we removed stopwords and stemmed the remaining terms using the Snowball stemmer <ref type="bibr" target="#b29">[30]</ref>.</p><p>Stylometric Features: Stylometry, the statistical analysis of variations in linguistic constructs, has been used with great success in distinguishing deceptive from truthful language <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b25">26]</ref>, and objective from subjective remarks <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b20">21]</ref>. Accordingly, we surmised that capturing stylistic variation will aid in the identification of check-worthy sentences as well, especially since they are typically expected to appear factual and objective.</p><p>In order to obtain shallow syntactic features from each sentence, we extracted the part-of-speech (POS) tags, the total number of tokens, and the number of tokens in past, present, and future tenses. We were able to infer the tense from the POS tags (e.g., both vbd and vbz are verb tags, but they indicate past and present tense, respectively). Additionally, we also extracted the number of negations in each sentence. More complex structural patterns of language, however, can only be captured by deep syntactic features. For that, we generated the constituency parse trees of all sentences, and selected clause-level and phrase-level tags. The number of words within the scope of each tag were included as the corresponding feature values. These tags, as defined in the Penn Treebank <ref type="bibr" target="#b1">[2]</ref>, are shown in Table <ref type="table" target="#tab_1">2</ref>. In addition to stylometry, the motivation behind using the number of words was to obtain a representation of the amount of information available under specific syntactic structures. Fig. <ref type="figure" target="#fig_0">1</ref> illustrates this point with the parse tree of a sentence from the training data that was labeled as check-worthy.</p><p>Semantic Features: We used the Stanford named entity recognizer (NER) <ref type="bibr" target="#b6">[7]</ref> to extract the number of named entities in a sentence. Additionally, we appended an extra feature for named entities of the type person.</p><p>Affective Features: We used the TextBlob <ref type="bibr" target="#b21">[22]</ref> library to train a naïve Bayes classifier on the pioneering movie review corpus for sentiment analysis <ref type="bibr" target="#b26">[27]</ref>, and thereby obtained a sentiment score for each sentence. In addition to overt sentiment, we also used the connotation of words in a sentence as features. For this, we employed Connotation WordNet <ref type="bibr" target="#b15">[16]</ref>, which assigns a (positive or negative) connotation score to each word. For every sentence, we queried this lexicon and retrieved the connotation score of its words. Finally, the overall connotation of the sentence was attributed simply to the mean of these scores.</p><p>Additionally, we also utilized lexicons that contain information about the subjective or objective nature of words <ref type="bibr" target="#b34">[35]</ref>, whether they directly indicate or are typically associated with language that indicates bias <ref type="bibr" target="#b30">[31]</ref>, and whether they are typically used to voice positive or negative opinions <ref type="bibr" target="#b14">[15]</ref>. For every sentence, we extracted the number of words in these categories (as defined by their scores in these lexicons), thus forming four new features: (i) subjectivity, (ii) direct bias, (iii) associated bias, and (iv) opinion.</p><p>Metadata Features: In addition to the syntactic and semantic features described above, we also included three binary non-linguistic features extracted from the training sample, indicating whether or not (i) the speaker's opponent is mentioned, (ii) the speaker is the anchor/moderator, or (iii) the sentence is immediately followed by intense reaction. The third feature is encoded in the training data as a 'system' reaction, as shown by the last sentence in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Discourse Features: All the above features were extracted without regards to the category (i.e., Debate and Speech). Since debates involve an interactive discourse structure where sentences are often formed as an immediate response to statements made by others, we include segments from the debates. We adopt the approach taken by Gencheva et al <ref type="bibr" target="#b8">[9]</ref> and regard a "segment" to be the maximal set of consecutive sentences by the same speaker. As features, we include the relative position of a sentence within its segment, and the number of sentences in the previous, current and subsequent segments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Feature Selection</head><p>The feature extraction processes described above yielded a very high-dimensional feature space. High-dimensional spaces, however, quickly lead to a decrease in the predictive power of models <ref type="bibr" target="#b31">[32]</ref>. Moreover, given the extreme class imbalance, classification in such a space is likely to ignore important features indicative of the minority class (in this case, the 'check-worthy' sentences).</p><p>To reduce the dimensionality, we applied a feature selection module using the scikit-learn library <ref type="bibr" target="#b28">[29]</ref>. As the first step, univariate feature selection was performed, and the 2,000 best features were selected based on χ 2 -test. Next, armed with the observation that linear predictive models with L1 loss yield sparse solutions and encourage the vanishing coefficients for weakly correlated features <ref type="bibr" target="#b24">[25]</ref>, we used a support vector machine (SVM) model with linear kernel and L1 regularization to further remove the relatively unimportant features. This step was first done on the entire training data, and then combined with repeated undersampling (without replacement) for the majority class. Each iteration of this undersampling process resulted in a small but balanced training sample. A L1-regularized SVM learner was trained on every sample generated in this manner, and features with vanishing coefficients were discarded. The cumulative effect of these feature selection steps was a reduction of the feature space to 2,655 and 2,404 dimensions for identification of check-worthy claims from debates and speeches, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Heuristics</head><p>Certain heuristics were introduced to override the scores assigned by the classification models. These rules differed slightly based on (i) the category, i.e., speech or debate, and (ii) whether or not the 'strict' heuristics were deployed. The strictness flag was introduced to control the threshold sentence size. When active, it would tend to discard more sentences.</p><p>These rules are specified in Algorithm 1. One particular rule required the identification of subjects in a sentence. To extract this information, we generated dependency parse trees of the sentences and counted the number of times any of the following dependency labels appeared: nsubj, csubj, nsubjpass, csubjpass, or xsubj. The first two indicate nominal and clausal subjects, respectively. The next two indicate nominal and clausal subjects in a passive clause, and the last label denotes a controlling subject, which relates an open clausal complement to its external clause.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Models</head><p>Our experiments comprised two supervised learning algorithms: support vector machines (SVM) and multilayer perceptrons (MLP). Additionally, we also built an ensemble model combing the two. In this section, we provide a description of these three models, along with their training processes.</p><p>For reasons described in Sec. 3.2, the SVM utilized a linear kernel with L1 regularization for feature selection. However, due to the propensity of the L1 loss function to miss optimal solutions, we used L2 loss in building the final model after completing feature selection. Our second model was the MLP. Here, we used two hidden layers with 100 units and 8 units in them, respectively. We used the hyperbolic tangent (tanh) as our activation function since it achieved better results when compared to rectified linear units (ReLU). Stochastic optimization was done with Adam <ref type="bibr" target="#b16">[17]</ref>. To avoid overfitting, we used L2-regularization in both Table <ref type="table">3</ref>. Results for the Check-Worthiness task of our submitted models: MLP was the primary submission, along with two contrastive runs, MLPstr and ENS (MLP with strict heuristics and the ensemble model, respectively). MLPnone shows the results of the MLP without any heuristics being applied. The primary evaluation metric was mean avg. precision (MAP). The mean reciprocal rank (MRR), mean R-precision (MRP), and mean precision at k (MP@k) are also shown.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>MAP</head><p>MRR MRP MP@1 MP@3 MP@5 MP@10 MP@20 MP@50 MLP SVM and MLP. Third, we built an ensemble model that combines SVM and MLP (without the strict heuristics). In this model, the final output score was a normalization (by standard deviation) of the results of SVM and MLP, and then computing the average. For all three models, class imbalance was a hindrance during the training process. To overcome that, we used ADASYN <ref type="bibr" target="#b13">[14]</ref>, an adaptive synthetic sampling algorithm for imbalanced learning. For model selection, we used 3-fold crossvalidation for debates, using two files for training and the remaining one for testing, to evaluate model performances and tune parameters. For speeches, we split the training sample into two halves (one file in each) for 2-fold crossvalidation. The evaluation script was provided by the task organizers, with the mean average precision (MAP) being the primary evaluation metric.</p><p>MLP without the strict heuristics demonstrated the best results during the training process, so this was submitted for the primary run. For the two contrastive runs, we submitted (i) MLP with strict heuristics, and (ii) the ensemble model without the strict heuristics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results and Analysis</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Empirical Results</head><p>The detailed performance of all three submissions we made is shown in Table <ref type="table">3</ref>. Even though MLP yielded the best training results without the strict heuristics, MLP str performed demonstrably better across multiple metrics on the test data. Our third model, the ensemble classifier, performed poorly in general compared to both MLP models. It did, however, achieve slightly better mean R-precision and mean precision at higher cutoffs (k = 10 and 50).</p><p>Without the inclusion of any heuristics, the performance of MLP dropped significantly. This was expected, since the heuristics were designed to address the flaws of the classifiers. This model was not among the submissions, but we include it here for comparison. The difference between MLP and MLP none quantifies the extent to which the rules help the supervised learners. Next, in Table <ref type="table" target="#tab_4">4</ref>, we present the comparison between the results obtained by all participants. This comparison was done only on the primary submission from each team. Our MLP model without the strict heuristics achieved the best MAP, MRR, and MRP scores. Further, it also outperformed the others in terms of correctly placing the check-worthy sentences at the very top of the ranked output list, as demonstrated by the mean precision at low values (k = 1 and 3).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Qualitative Analysis</head><p>Identifying check-worthy sentences is a difficult and novel task, and even the best model suffered from misclassification errors. Upon analyzing such mistakes made by the MLP models, we were able to discern a few reasons.</p><p>First, tense plays a logical role in check-worthiness, since future actions cannot be verified. However, the part-of-speech tagging often confuses the future tense with the present continuous (e.g., "We're cutting taxes."). Second, we observed that anecdotal stories are often highly prioritized as check-worthy, while they are not. These sentences are usually complex, with a lot of content, which makes it easy for the model to conflate them with other complex sentences pertaining to real events deemed check-worthy. Third, the presence of duplicate sentences in the data means that a misclassification gets amplified, while the presence of very similar sentences with different labels likely makes the feature selection stage discard potentially useful features.</p><p>At a more abstract level, rhetorical figures of speech play a critical role. They often break the structures associated with standard sentence formation. Several sentences that were misclassified exhibited constructs such as scesis onomaton, where words or phrases with nearly equivalent meaning are repeated. We conjecture that this makes the model falsely believe that there is more informational content in the sentence. Such figures of speech become even harder to handle when they occur across multiple speakers in debates. The conversational aspect of debates also causes another problem: quite a few sentences are short, and in isolation, would perhaps not be check-worthy. However, as a response to things mentioned earlier in the debate, they are.</p><p>Another complex issue leading to misclassification is the use of sentence fragments. This is sparingly used for dramatic effect in literature, but was seen with alarming frequency in the political debates due to the prevalence of illformed or partly-formed sentences stopping and then giving way to another sentence. In some cases, the fragments are portions of the sentence that the speaker repeats. An example of such a fragment is the sentence "Ambassador Stevens -Ambassador Stevens sent 600 requests for help.", where the phrase "Ambassador Stevens" is repeated.</p><p>A proper approach to deal with these hurdles is a complex matter in and by itself. We believe that our features are better suited for written language than speech or debate transcripts. In the presence of significantly more labeled data for check-worthiness, ablation studies that remove such sentences could provide empirical evidence of this intuition.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion and Future Work</head><p>We developed a hybrid system that combines a few rules with supervised learning to detect check-worthy sentences in political debates and speeches. To tackle the severity of class imbalance, our development also included a sophisticated feature selection process and special sampling methods. Our primary model achieved the best results among all participants over multiple performance metrics.</p><p>This work opens up several intriguing possibilities for future research in the field of fact-checking. First, we intend to study in greater details the linguistic forms of informational content. Shallow syntax has been explored to understand this aspect of language in sociolinguistics, and some work has even looked into deep syntactic features. This approach has, however, not yet been applied to identifying check-worthy sentences. Furthermore, more complex neural network structures need to be thoroughly investigated. Along this line, we will be investigating deep learning models with feedback control. A stringent and focused work on these issues will empower journalists and citizens alike to be better informed and more cognizant of false claims permeating news and social media now. To that end, we also need complementary advances in related areas like natural language querying, crowdsourcing, source identification, and social network analysis.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The constituency parse tree of a check-worthy sentence from the training data: "President Bush said we would leave Iraq at the end of 2011." The size of the subtree under the subordinate clause (sbar) is representative of the amount of information available provided about the action 'said' undertaken by the entity 'President Bush'.</figDesc><graphic coords="6,193.75,115.84,224.78,173.38" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Labeled sentence examples from political debates provided as training data. Check-worthy sentences are labeled 1, and others are labeled 0. Audience reaction and other background noise is encoded as "SYSTEM"-generated.</figDesc><table><row><cell>Speaker</cell><cell>Sentence</cell><cell>Label</cell></row><row><cell>HOLT</cell><cell>I'm Lester Holt, anchor of "NBC Nightly News."</cell><cell>0</cell></row><row><cell>HOLT</cell><cell>I want to welcome you to the first presidential debate.</cell><cell>0</cell></row><row><cell>TRUMP</cell><cell>Our jobs are fleeing the country.</cell><cell>0</cell></row><row><cell>TRUMP</cell><cell>Thousands of jobs leaving Michigan, leaving Ohio.</cell><cell>1</cell></row><row><cell>CLINTON</cell><cell>Donald thinks that climate change is a hoax</cell><cell>1</cell></row><row><cell></cell><cell>perpetrated by the Chinese.</cell><cell></cell></row><row><cell>SYSTEM</cell><cell>(applause)</cell><cell>0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Constituent tags from the Penn Treebank.</figDesc><table><row><cell>Clause-Level</cell><cell>SBAR, SBARQ, SINV, SQ, S</cell></row><row><cell>Phrase-Level</cell><cell>ADJP, ADVP, CONJP, FRAG, INTJ, LST, NAC, NP, NX,</cell></row><row><cell></cell><cell>PP, PRN, PRT, QP, RRC, UCP, VP, WHADJP, WHAVP,</cell></row><row><cell></cell><cell>WHNP, WHPP, X</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>Algorithm 1 Heuristics for assigning the check-worthiness score w(•) to sentences.</figDesc><table><row><cell>Require: category ∈ {speech, debate},</cell><cell>if Sspeaker is system then</cell></row><row><cell>strict mode ∈ {true, false}, sentence S.</cell><cell>w(s) ← 10 −8</cell></row><row><cell></cell><cell>end if</cell></row><row><cell>min token count ← 0</cell><cell>if Snumber of tokens &lt; min token count</cell></row><row><cell>if category is speech then</cell><cell>then</cell></row><row><cell>if strict mode then</cell><cell>w(s) ← 10 −8</cell></row><row><cell>min token count ← 10</cell><cell>end if</cell></row><row><cell>else</cell><cell>if S contains "thank you" then</cell></row><row><cell>min token count ← 8</cell><cell>w(s) ← 10 −8</cell></row><row><cell>end if</cell><cell>end if</cell></row><row><cell>else</cell><cell>if Snumber of subjects &lt; 1 then</cell></row><row><cell>if strict mode then</cell><cell>if category is speech then</cell></row><row><cell>min token count ← 7</cell><cell>w(s) ← 10 −8</cell></row><row><cell>else</cell><cell>else if S contains "?" then</cell></row><row><cell>min token count ← 5</cell><cell>w(s) ← 10 −8</cell></row><row><cell>end if</cell><cell>end if</cell></row><row><cell>end if</cell><cell>end if</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 .</head><label>4</label><figDesc>Results from the primary submissions from all participants. We participated under the name Prise de Fer. The best results for each metric is shown in bold.</figDesc><table><row><cell>TEAM</cell><cell>MAP</cell><cell>MRR</cell><cell>MRP</cell><cell cols="6">MP@1 MP@3 MP@5 MP@10 MP@20 MP@50</cell></row><row><cell>Prise de Fer</cell><cell cols="5">0.1332 0.4965 0.1352 0.4286 0.2857</cell><cell>0.2000</cell><cell>0.1429</cell><cell>0.1571</cell><cell>0.1200</cell></row><row><cell>Copenhagen</cell><cell>0.1152</cell><cell>0.3159</cell><cell>0.1100</cell><cell>0.1429</cell><cell>0.1429</cell><cell>0.1143</cell><cell>0.1286</cell><cell>0.1286</cell><cell>0.1257</cell></row><row><cell>UPV-INAOE</cell><cell>0.1130</cell><cell>0.4615</cell><cell>0.1315</cell><cell>0.2857</cell><cell>0.2381</cell><cell>0.3143</cell><cell>0.2286</cell><cell>0.1214</cell><cell>0.0866</cell></row><row><cell>bigIR</cell><cell>0.1120</cell><cell>0.2621</cell><cell>0.1165</cell><cell>0.0000</cell><cell>0.1429</cell><cell>0.1143</cell><cell>0.1143</cell><cell>0.1000</cell><cell>0.1114</cell></row><row><cell>fragarach</cell><cell>0.0812</cell><cell>0.4477</cell><cell>0.1217</cell><cell>0.2857</cell><cell>0.1905</cell><cell>0.2000</cell><cell>0.1571</cell><cell>0.1071</cell><cell>0.0743</cell></row><row><cell>blue</cell><cell>0.0801</cell><cell>0.2459</cell><cell>0.0576</cell><cell>0.1429</cell><cell>0.0952</cell><cell>0.0571</cell><cell>0.0571</cell><cell>0.0857</cell><cell>0.0600</cell></row><row><cell>RNCC</cell><cell>0.0632</cell><cell>0.3755</cell><cell>0.0639</cell><cell>0.2857</cell><cell>0.1429</cell><cell>0.1143</cell><cell>0.0571</cell><cell>0.0571</cell><cell>0.0486</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">The lab task provided all seven files together, without this categorization into speeches and debates. We, however, chose to treat these differently since language use is very different in these two scenarios: debates consist of the interactive statements made by the candidates and the moderator, while speeches only have a single speaker, and there is no two-sided conversational structure.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">The provided training sample included two speeches, and both were by Donald Trump. As a result, for the purpose of this task, a single sub-dataset was created. The approach is independent of the speaker and the number of speakers, however.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">Available at https://code.google.com/archive/p/word2vec/.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgment: This work was supported in part by the U.S. National Science Foundation (NSF) under the award SES-1834597.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2018 CheckThat! Lab on Automatic Identification and Verification of Political Claims, Task 1: Check-Worthiness</title>
		<author>
			<persName><forename type="first">P</forename><surname>Atanasova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kyuchukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<ptr target="org" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2018 Working Notes. Working Notes of CLEF 2018 -Conference and Labs of the Evaluation Forum. CEUR Workshop Proceedings, CEUR-WS</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Nie</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<meeting><address><addrLine>Avignon, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-09">September 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Bracketing Guidelines for Treebank II Style Penn Treebank Project</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ferguson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Macintyre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tredinnick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Marcinkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schasberger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">University of Pennsylvania</title>
		<imprint>
			<biblScope unit="volume">97</biblScope>
			<biblScope unit="page">100</biblScope>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Blogs, Twitter, and breaking news: the produsage of citizen journalism</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bruns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Highfield</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Produsing Theory in a Digital World: The Intersection of Audiences and Production in Contemporary Theory</title>
				<imprint>
			<publisher>Peter Lang Publishing Inc</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">80</biblScope>
			<biblScope unit="page" from="15" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Cazalens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lamarre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leblay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Manolescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tannier</surname></persName>
		</author>
		<title level="m">Journalism, Misinformation and Fact Checking&quot; alternate paper track of&quot; The Web Conference</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note>A content management perspective on fact-checking</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Computational Journalism: A Call to Arms to Database Researchers</title>
		<author>
			<persName><forename type="first">S</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Innovative Data Systems Research. CIDR &apos;11</title>
				<meeting><address><addrLine>Asilomar, California, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Syntactic Stylometry for Deception Detection</title>
		<author>
			<persName><forename type="first">S</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Banerjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 50 th Annual Meeting of the Association for Computational Linguistics: Short Papers</title>
				<meeting>the 50 th Annual Meeting of the Association for Computational Linguistics: Short Papers</meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="171" to="175" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Incorporating non-local information into information extraction systems by Gibbs sampling</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Finkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Grenager</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 43 rd Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 43 rd Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="363" to="370" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The promise of computational journalism</title>
		<author>
			<persName><forename type="first">T</forename><surname>Flew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Spurgeon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Daniel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Swift</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journalism Practice</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="157" to="171" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A contextaware approach for detecting worth-checking claims in political debates</title>
		<author>
			<persName><forename type="first">P</forename><surname>Gencheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Koychev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference Recent Advances in Natural Language Processing</title>
				<meeting>the International Conference Recent Advances in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="267" to="276" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Social news, citizen journalism and democracy</title>
		<author>
			<persName><forename type="first">L</forename><surname>Goode</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">New media &amp; society</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1287" to="1305" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Distributional Structure</title>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">S</forename><surname>Harris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Word</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="146" to="162" />
			<date type="published" when="1954">1954</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Detecting check-worthy factual claims in presidential debates</title>
		<author>
			<persName><forename type="first">N</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tremayne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 24 th ACM International Conference on Information and Knowledge Management</title>
				<meeting>the 24 th ACM International Conference on Information and Knowledge Management</meeting>
		<imprint>
			<publisher>CIKM</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="1835" to="1838" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">ClaimBuster: The First-ever End-to-end Fact-checking System</title>
		<author>
			<persName><forename type="first">N</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Arslan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Caraballo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jimenez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gawsane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hasan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joseph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kulkarni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Nayak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the VLDB Endowment</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1945" to="1948" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">ADASYN: Adaptive Synthetic Sampling Approach for Imbalanced Learning</title>
		<author>
			<persName><forename type="first">H</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">A</forename><surname>Garcia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Joint Conference on Neural Networks (IJCNN)</title>
				<meeting>the IEEE Joint Conference on Neural Networks (IJCNN)</meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="1322" to="1328" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Mining and Summarizing Customer Reviews</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10 th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 10 th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="168" to="177" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">ConnotationWordNet: Learning Connotation over the Word+Sense Network</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Akoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 52 nd Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 52 nd Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2014-06">June 2014</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1544" to="1554" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Adam: A method for stochastic optimization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6980</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Varieties of Confirmation Bias</title>
		<author>
			<persName><forename type="first">J</forename><surname>Klayman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychology of learning and motivation</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="385" to="418" />
			<date type="published" when="1995">1995</date>
			<publisher>Elsevier</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Separating Fact from Fear: Tracking Flu Infections on Twitter</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lamb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Paul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dredze</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="789" to="795" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Towards a text analysis system for political debates</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">T</forename><surname>Vu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Blessing</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10 th SIGHUM Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities</title>
				<meeting>the 10 th SIGHUM Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="134" to="139" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Objectivity Classification in Online Media</title>
		<author>
			<persName><forename type="first">E</forename><surname>Lex</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Juffinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Granitzer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st ACM Conference on Hypertext and Hypermedia</title>
				<meeting>the 21st ACM Conference on Hypertext and Hypermedia</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="293" to="294" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">TextBlob: Simplified Text Processing</title>
		<author>
			<persName><forename type="first">S</forename><surname>Loria</surname></persName>
		</author>
		<ptr target="http://textblob.readthedocs.org/en/dev/(2014" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Efficient Estimation of Word Representations in Vector Space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1301.3781</idno>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2018 Lab on Automatic Identification and Verification of Claims in Political Debates</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gencheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kyuchukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2018 -Conference and Labs of the Evaluation Forum. CLEF &apos;18</title>
				<meeting><address><addrLine>Avignon, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-09">September 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Feature selection, l 1 vs. l 2 regularization, and rotational invariance</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the twenty-first international conference on Machine learning</title>
				<meeting>the twenty-first international conference on Machine learning</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page">78</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Finding deceptive opinion spam by any stretch of the imagination</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cardie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">T</forename><surname>Hancock</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 49 th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 49 th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="309" to="319" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Thumbs up?: sentiment classification using machine learning techniques</title>
		<author>
			<persName><forename type="first">B</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vaithyanathan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACL-02 conference on Empirical methods in natural language processing-Volume 10</title>
				<meeting>the ACL-02 conference on Empirical methods in natural language processing-Volume 10</meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="79" to="86" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">TATHYA: A Multi-Classifier System for Detecting Check-Worthy Statements in Political Debates</title>
		<author>
			<persName><forename type="first">A</forename><surname>Patwari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Goldwasser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bagchi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 26 th ACM International Conference on Information and Knowledge Management</title>
				<meeting>the 26 th ACM International Conference on Information and Knowledge Management</meeting>
		<imprint>
			<publisher>CIKM</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">Snowball: A Language for Stemming Algorithms</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Porter</surname></persName>
		</author>
		<ptr target="http://snowball.tartarus.org/texts/introduction.html" />
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Linguistic Models for Analyzing and Detecting Biased Language</title>
		<author>
			<persName><forename type="first">M</forename><surname>Recasens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Danescu-Niculescu-Mizil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 51st Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1650" to="1659" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">A Problem of Dimensionality: A Simple Example</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Trunk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="306" to="307" />
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Fact Checking: Task definition and dataset construction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vlachos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACL 2014 Workshop on Language Technologies and Computational Social Science</title>
				<meeting>the ACL 2014 Workshop on Language Technologies and Computational Social Science</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="18" to="22" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Werner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Tankard</surname><genName>Jr</genName></persName>
		</author>
		<title level="m">Communication theories: Origins, methods and uses in the mass media</title>
				<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wilson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wiebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hoffmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing</title>
				<meeting>the Conference on Human Language Technology and Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="347" to="354" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Toward computational fact-checking</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the VLDB Endowment</title>
				<meeting>the VLDB Endowment</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="589" to="600" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
