<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">bigIR at CLEF 2018: Detection and Verification of Check-Worthy Political Claims</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Khaled</forename><surname>Yasser</surname></persName>
							<email>khaled.yasser@qu.edu.qa</email>
							<affiliation key="aff0">
								<orgName type="institution">Qatar University</orgName>
								<address>
									<settlement>Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mucahid</forename><surname>Kutlu</surname></persName>
							<email>mucahidkutlu@qu.edu.qa</email>
							<affiliation key="aff0">
								<orgName type="institution">Qatar University</orgName>
								<address>
									<settlement>Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tamer</forename><surname>Elsayed</surname></persName>
							<email>telsayed@qu.edu.qa</email>
							<affiliation key="aff0">
								<orgName type="institution">Qatar University</orgName>
								<address>
									<settlement>Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">bigIR at CLEF 2018: Detection and Verification of Check-Worthy Political Claims</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">1797DD16B589609BC49FD1765BD91D5F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T02:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Fact-Checking</term>
					<term>Check-Worthiness</term>
					<term>Veracity Prediction</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>With the enormous amount of misinformation spread over the Internet, manual fact-checking is no longer feasible to prevent its negative impact. There is an urgent need for automated systems that can make fact-checking process faster and effectively detect the veracity of claims. In this paper, we present our participation in the two tasks of CLEF-2018 CheckThat! Lab. To rank claims based on their check-worthiness (Task 1), we propose a learning-to-rank approach with features extracted by natural language processing such as named entity recognition and sentiment analysis. For veracity prediction (Task 2), we propose using an external Web search engine to retrieve potentiallyrelevant Web pages and extract features from relevant segments of those pages to predict the veracity. In the official evaluation, our best performing runs for Task 1 are ranked 4 th (out of 16 runs from 8 teams) and 1 st (out of 5 runs from 2 teams) over the English and Arabic datasets respectively, while our best performing run for Task 2 is ranked 6 th (out of 10 runs from 5 teams) over the English datasets.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In the wake of the widespread of misinformation on the Internet and in the news, there emerged a need to combat this phenomenon as effectively as possible. However, despite the advent of technology, human fact-checkers could not keep up with the pace in which misinformation is perpetuated and spread. As a result, there is a need to develop automated systems to assist and help combat false claims and misinformation.</p><p>CLEF-2018 CheckThat! Lab <ref type="bibr" target="#b8">[9]</ref> introduced two tasks which address two important aspects of fact-checking systems. The main goal of the first task <ref type="bibr" target="#b0">[1]</ref> is to detect the check-worthy claims in political debates and prioritize them based on their check-worthiness. This resultant system has two benefits: <ref type="bibr" target="#b0">(1)</ref> filtering the claims to be automatically fact-checked, and (2) helping human fact-checkers prioritize the claims to be fact-checked to focus on the most important ones. The second task <ref type="bibr" target="#b1">[2]</ref> aims at predicting the veracity of the claims automatically, which is the ultimate goal of fact-checking systems. CheckThat! Lab released both English and Arabic datasets for the two tasks. In this paper, we present our approaches to tackle these two important problems in the context of the lab.</p><p>For the check-worthiness task, we use a learning-to-rank approach with features extracted for each sentence in the debates. The features include word embeddings, types of named entities, part-of-speech tags, and sentiment and topic of sentences. We evaluated the impact of each feature group on training data and eventually submitted (for the test phase) runs on 3 models where each model has different sets of feature groups. Our best performing model that uses only features of named entities, sentiment and topic of sentences is ranked 4 th in the official evaluation. For the Arabic dataset, we automatically translated the dataset into English and used the same models. Our best performing model ranked 1 st in the official evaluation, but only 2 groups participated in the Arabic data challenge.</p><p>For the factuality task, we first retrieve some potentially-relevant Web pages using a commercial Web search engine, excluding the pages not allowed (by the lab organizers) for the task. Next, we detect the relevant segments within the pages and use them to extract features for each claim. Our features include percentages of sentences confirming the claim and contradicting with the claim, and also stances of relevant segments. Our best performing run is ranked 6 th in the official evaluation for the English dataset, but we did not participate in the Arabic data challenge in this task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task 1: Check-Worthiness</head><p>In Task 1 <ref type="bibr" target="#b0">[1]</ref>, the goal is to rank sentences made in presidential debates according to their likelihood of being check-worthy. In this section, we describe our approach and present the evaluation results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Proposed Approach</head><p>Prioritizing the claims based on their check-worthiness is a ranking problem. Therefore, our approach relies on using learning-to-rank (L2R) methods. There are two versions of the datasets, one in English, and another in Arabic. First, we explain our approach for the English dataset, then how we modify our approach for the Arabic dataset.</p><p>Learning-to-Rank Model. We propose an L2R model for this task because the goal is to rank claims based on their check-worthiness instead of classifying them. We focus on pairwise and point-wise L2R models since we do not have a list of queries. We use RankLib L2R toolkit<ref type="foot" target="#foot_0">1</ref> in our implementation. In our experiments on the training data, we observed that the MART algorithm <ref type="bibr" target="#b3">[4]</ref> outperforms other L2R algorithms we tried (i.e., RankBoost and RankNet). Therefore we opted for the MART algorithm in the testing phase.</p><p>Features. The sentences in the dataset can be a full or half sentence. Because the data is extracted from debates, there exist sentences used in daily language, sentences without any verb because of being interrupted by another person, or sentences that are just a continuation of a previously interrupted sentence. While this kind of challenge suggests to consider sentences before and/or after each sentence in feature extraction, we chose to ignore these contextual data and consider only the individual sentences because the features extracted from the context would overlap with other statements, which could potentially introduce noise. The features we extract for each sentence can be divided into 5 categories:</p><p>(1) vector representation of sentences, (2) named entities, (3) part-of-speech tags, (4) sentiment, and (5) topic features. We explain each of these features next.</p><p>-Vector Representation of Sentence: We use Word2Vec to represent sentences. Due to the small sample set in our training data, we use embeddingbased vector representation instead of term-based representation. This allows us to capture similar sentences, and not just ones with the exact terms used.</p><p>It also has lower dimensions as opposed to that of a term-based representation. We used a model pre-trained on Google News<ref type="foot" target="#foot_1">2</ref> where each vector has 300 components, and represented each sentence by the average vector of the vectors of its terms. -Types of Named Entities: Check-worthy claims usually contain some named entities such as organizations, countries, and people. However, not all types of entities will be helpful for identifying check-worthy claims. For example, a claim which mentions the name of an international company is more likely to be check-worthy than a claim which mentions the name of a local music group, both of which are valid named entity types. We represent the types of named entities as an n-dimensional vector where each dimension corresponds to a type. The vector contains binary features, reflecting the existence or absence of a certain entity type in a sentence. In order to detect the named entity types, we use IBM Watson API for Natural Language Understanding<ref type="foot" target="#foot_2">3</ref> which yields a vector of 26 features. Some of the available entity tags are: person, organization, country, and geographical entity. -Part-of-Speech (POS) Tags: There can be structural aspects of sentences that can help distinguish check-worthiness. For example, a sentence written in future tense is less likely to be check-worthy than a sentence written in past tense. Therefore, we use POS tags to capture the sentence structure of claims. We represent POS tags as a vector of binary features where each feature corresponds to the existence or absence of a certain tag. We use Stanford CoreNLP <ref type="bibr" target="#b6">[7]</ref> to extract the POS tags. The POS feature vector consists of 36 features. -Sentiment: We use sentiment of sentences (i.e., whether the sentence is presenting a positive, negative, or neutral attitude) as a feature. We extract sentiment labels using Stanford CoreNLP <ref type="bibr" target="#b6">[7]</ref>, but we collapse different grades of positive and negative labels (e.g., very positive and very negative) to positive and negative, respectively. -Topic: The topic of the sentence can indicate whether it is worth being checked or not. For example, a sentence about the use of nuclear weapons is more check-worthy than one about the release date of a movie. Therefore, we extracted topics of sentences using IBM Watson API in which a sentence could be classified into multiple topics where each topic could be fine-grained up to three levels. In our implementation, we consider only topics which have a confidence score of 0.5 or higher and used only the first two levels; after manual inspection of the topics available we found the third level to be too fine-grained for such a small dataset. The topics, just like types of named entities and POS tags, are represented as a vector of binary features but with a dimensionality of 348.</p><p>Feature Selection. After constructing the feature vectors for all sentences, we performed two-tier feature selection to choose the most discriminant features. The first tier is group-level selection, in which we evaluate the impact of each feature group previously mentioned in a leave-one-out fashion. For example, we test a model with and without topic features to see their impact on the results. The first tier was done manually by trying different possible combinations of feature groups (See Section 2.2). The second tier adopts a variance-threshold feature selection where we removed features which have variance less than a certain threshold across the samples. We used a threshold of 0 to eliminate features which have the same value across all samples.</p><p>Handling Class Imbalance. We observed that the training dataset given for Task 1 is highly imbalanced such that only 3% of the statements have been judged as check-worthy. This could potentially cause problems on the test data due to lack of enough samples for check-worthy claims. Therefore, we oversample the positive samples by duplication. In our experiments with the training data, we tried multiple oversampling values and observed that an oversample factor of 400% is the most ideal one for the training data.</p><p>Arabic. All the steps we described so far have been applied on the English dataset. One of the main challenge for Arabic dataset is that there is no suitable NLP tools that can be used to extract features. Therefore, we opted for using machine translation methods. Mohammad et al. <ref type="bibr" target="#b7">[8]</ref> discussed the effect of machine translation on sentiment analysis and showed that models trained on a different language yield results comparable to those trained on the same language. Therefore, we first translate the Arabic dataset to English using Google Translate API <ref type="foot" target="#foot_3">4</ref> and then apply the same model trained on the English one.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Evaluation Results</head><p>In order to pick the models to use in the test phase, we performed k -fold cross validation for evaluating different models. Testing the models on data from the same debates they are trained on might not give proper insights on how they would perform on the unseen data; they could share similar topics, talk about the same entities, or have similar syntactic styles. Therefore, we set the folds such that each has the data from a separate debate.</p><p>In our unreported initial experiments, we observed that MART outperforms other L2R algorithms when the maximum number of trees and learning rate are set to 100 and 0.1 respectively. Subsequently, we evaluated the impact of each feature group using the MART algorithm while also changing the number of leaves parameter. The values for the number of leaves parameter we tried are from 5 to 12 inclusive. Table <ref type="table" target="#tab_1">2</ref> As shown in Table <ref type="table" target="#tab_1">2</ref>.2, the best performing model is the one trained without using Word2Vec vectors as features using trees with only 5 leaves. Aside from Word2Vec, models trained without POS features also outperform the models that do not use a particular set of features. We select the top 3 models (shown in bold in the table) for the submission of both English and Arabic datasets without picking two models with the same feature groups.</p><p>In the official evaluation for the English dataset, our selected models All -{Word2Vec}, All -{POS tags, Word2Vec}, and All -{POS tags} achieved MAP values of 0.1120 (10 th in the ranking), 0.1319 (4 th in the ranking), and 0.1117 (11 th in the ranking) respectively. As on the Arabic dataset, our models achieved 0.0899 (4 th in the ranking), 0.1498 (1 st in the ranking), and 0.0962 (3 rd in the ranking) MAP scores respectively. That clearly shows the second model (i.e., All -{POS tags, Word2Vec}) outperforms the other two models in both datasets. Interestingly, it performed better on the Arabic dataset than it did on the English one (0.1498 vs. 0.1319).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Task 2: Factuality</head><p>After finding which claims are more important to be checked, Task 2 <ref type="bibr" target="#b1">[2]</ref> focuses on the second stage of automated fact-checking which is detecting the veracity of check-worthy claims.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Proposed Approach</head><p>We approach this task as a classification problem. For a given claim, we first find potentially-relevant Web pages using an external Web search engine. Then we detect the relevant segments in each of those pages and extract features from these segments. Finally, we predict the veracity of the claim using a learning model based on the extracted features. We discuss our approach in detail next.</p><p>Web Search. The first step of our approach is retrieving search results for the claim. We follow the approach described by Karadzhov et al. <ref type="bibr" target="#b4">[5]</ref> to generate a query from a given claim. We use Google Custom Search<ref type="foot" target="#foot_4">5</ref> to retrieve the results but with custom settings to filter out websites which are not allowed for the task. We retrieve 10 results for each claim.</p><p>Relevant Segments Detection. The goal of this step is to find which parts of a Web page are relevant to the claim of interest. We compute the cosine similarity between the Word2Vec vectors of the claim and each sentence in the page. We consider a sentence as relevant if the cosine similarity score is higher than 0.5. Next, for each relevant sentence, we add a sentence before and a sentence after in order to capture the contextual information. We call these three-sentence structures relevant segments.</p><p>Features. We extract features from each relevant segment. We first explain the features and how they are extracted, then we explain how we aggregate the features from different pages. The features are as follows:</p><p>-Stance Detection: Stance detection is the process of finding whether a piece of text is for, against, or unrelated to another piece of text. Following the Fake News Challenge<ref type="foot" target="#foot_5">6</ref> , where the challenge was to check if an article supports or denies a claim using stance detection, we incorporate stance detection into our method. We use the implementation provided by <ref type="bibr" target="#b9">[10]</ref> and train it on the entire Emergent dataset <ref type="bibr" target="#b2">[3]</ref>. We first classify each relevant segment in all pages separately and then calculate the percentage of each label, yielding a feature vector of size 3. -Contradiction in Predicates: A stance can be a subjective statement regarding a claim. For instance, let the claim be "He bought weapons" and a sentence extracted from a document be "I do not want to believe that he bought weapons." While the sentence is against the claim, it does not state whether the claim is actually true or not. On the other hand, consider the following sentence: "He forgot to buy weapons". The predicate of this claim ("buy") contradicts the one in this sentence. In this feature, we try to detect if the predicate of a relevant segment contradicts with the predicate of the given claim. In order to extract such relation, we use TruthTeller <ref type="bibr" target="#b5">[6]</ref> in combination with WordNet <ref type="foot" target="#foot_6">7</ref> .</p><p>In particular, we first generate a relation matrix, R, which represents the relation between two sentences. R i,j indicates the relation between term i in the first sentence and term j in the other, where the relation can be either synonyms, antonyms, or unrelated. We compare the lemmas of the terms and assign the label synonyms if the two terms are the same. Furthermore, we classify each term in a sentence using TruthTeller where a term can either be positive (i.e., the term is affirmed), negative (i.e., the term is negated in some way), uncertain (i.e., there is a doubt around the term), or unknown (i.e., a decision could not be made). We detect the relative truth of term j in sentence B based on a term i in sentence A, given that they have a relation, using the following rules:</p><formula xml:id="formula_0">• If R i,j = synonyms and T (A i ) = T (B j ) → confirms • If R i,j = synonyms and T (A i ) = T (B j ) → contradicts • If R i,j = antonyms and T (A i ) = T (B j ) → contradicts • If R i,j</formula><p>= antonyms and T (A i ) = T (B j ) → confirms Figure <ref type="figure" target="#fig_1">1 and 2</ref> show an example of a relation matrix and truth predicates for the sentences "He bought weapons" and "He forgot to buy weapons" respectively. The terms "buy" and "bought" share the same lemma; their relation is set to synonyms. The terms "buy" and "bought" are synonyms with opposite truth values, positive and negative, respectively. Following the aforementioned rules, the two sentences contradict one another. The feature vector for predicates consists of two features, one for the percentage of confirming predicates, and one for the percentage of contradicting ones. We do exclude uncertain and unknown truth values to simplify the process since they do not offer any value.</p><p>The results of both stages, stance detection and contradiction in predicates, from all pages are aggregated into a single vector of features. This final vec-  tor carries the following 5 features: the percentage of segments supporting the claim, the percentage of segments which are against the claim, the percentage of segments unrelated to the claim, the percentage of sentences with contradicting predicates, and the percentage of sentences with confirming predicates.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Evaluation Results</head><p>The evaluation process for Task 2 is similar to that of Task 1. We performed k -fold cross-validation where each fold corresponds to the claims from one debate. The optimization and evaluation for Task 2 was done on model prediction accuracy. We tried three common classifiers for this problem and performed grid search optimization for each classifier on all of its parameters. The results are reported in Table <ref type="table" target="#tab_1">2</ref>. Although stance is more common, the results in Table <ref type="table" target="#tab_1">2</ref> show that using contradiction in predicate, by itself or in addition to stance, has the potential to improve classification accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Features</head><p>The selected models, highlighted in bold in Table <ref type="table" target="#tab_1">2</ref>, are SVM with contradictionin-predicates features, logistic regression with all features, and random forest with stance features. The metric in the official evaluation was Mean Squared Error (MSE), and our models received scores of 0.9640, 0.9640, and 0.9425, respectively. The first two models landed the last place, while the last one landed the 6 th place (out of 10).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>In this paper, we present our methods for the tasks of CLEF-2018 CheckThat! Lab. For the task of detecting check-worthy claims, we proposed a learningto-rank based approach which uses natural language processing methods for extracting the features. Our best performing model achieved the 4 th place on the English dataset. The same model, when used in conjunction with machine translation, got the 1 st place in the evaluation on the Arabic datasets.</p><p>Regarding the task of verifying the check-worthy claims, we proposed a system which uses an external Web search engine to collect evidence and then predicts the veracity of the claims by detecting the stance of relevant statements and contradicting or confirming sentences in the retrieved pages. Our best performing model got 6 th place. However, we observed promising results for the contradiction in predicates feature, suggesting that it can be a useful feature for fact-checking with more enhanced methods. However, it needs further experiments on larger datasets.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The relation matrix R for the sentences A = "He bought weapons" and B = "He forgot to buy weapons". Each cell Ri,j represents the relation between term i in A and term j in B.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig.2. The truth vectors for the sentences A = "He bought weapons" and B = "He forgot to buy weapons" .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>.2 shows the Average Precision (AP) score of the two best-performing models for each group of features on the English dataset. Results of training MART models with different groups of features and number of leaves for Task 1. The maximum number of trees in the training stage is set to 100. The ones in bold represent the models selected for the test phase.</figDesc><table><row><cell>Features</cell><cell cols="2">Number of leaves AP on test</cell></row><row><cell>All</cell><cell>10</cell><cell>0.196</cell></row><row><cell>All</cell><cell>12</cell><cell>0.195</cell></row><row><cell>All -{entity types}</cell><cell>6</cell><cell>0.241</cell></row><row><cell>All -{entity types}</cell><cell>7</cell><cell>0.250</cell></row><row><cell>All -{topics}</cell><cell>5</cell><cell>0.247</cell></row><row><cell>All -{topics}</cell><cell>6</cell><cell>0.241</cell></row><row><cell>All -{POS tags}</cell><cell>5</cell><cell>0.332</cell></row><row><cell>All -{POS tags}</cell><cell>6</cell><cell>0.281</cell></row><row><cell>All -{sentiment}</cell><cell>5</cell><cell>0.257</cell></row><row><cell>All -{sentiment}</cell><cell>8</cell><cell>0.216</cell></row><row><cell>All -{Word2Vec}</cell><cell>5</cell><cell>0.420</cell></row><row><cell>All -{Word2Vec}</cell><cell>6</cell><cell>0.337</cell></row><row><cell>All -{POS tags, Word2Vec}</cell><cell>5</cell><cell>0.399</cell></row><row><cell>All -{POS tags, Word2Vec}</cell><cell>6</cell><cell>0.347</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Results of training multiple models with stance and contradiction in predicates features for Task 2.</figDesc><table><row><cell>Classifier</cell><cell>Accuracy</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://sourceforge.net/p/lemur/wiki/RankLib/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://code.google.com/archive/p/word2vec/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://www.ibm.com/watson/services/natural-language-understanding/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://cloud.google.com/translate/docs/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://cse.google.com/cse/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">http://www.fakenewschallenge.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://wordnet.princeton.edu/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was made possible by NPRP grant# NPRP 7-1313-1-245 and NPRP grant# 7-1330-2-483 from the Qatar National Research Fund (a member of Qatar Foundation). Statements made herein are solely the responsibility of the authors.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2018 CheckThat! Lab on automatic identification and verification of political claims</title>
		<author>
			<persName><forename type="first">P</forename><surname>Atanasova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kyuchukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Task 1: Check-worthiness</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2018 CheckThat! Lab on automatic identification and verification of political claims</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Atanasova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kyuchukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Task</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<publisher>Factuality</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Emergent: a novel data-set for stance classification</title>
		<author>
			<persName><forename type="first">W</forename><surname>Ferreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vlachos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies</title>
				<meeting>the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1163" to="1168" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Greedy function approximation: a gradient boosting machine</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Annals of statistics</title>
				<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="1189" to="1232" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Fully automated fact checking using external sources</title>
		<author>
			<persName><forename type="first">G</forename><surname>Karadzhov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barron-Cedeno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Koychev</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1710.00341</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Truthteller: Annotating predicate truth</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lotan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stern</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Dagan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="752" to="757" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The Stanford CoreNLP natural language processing toolkit</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Surdeanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Finkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Bethard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mcclosky</surname></persName>
		</author>
		<ptr target="http://www.aclweb.org/anthology/P/P14/P14-5010" />
	</analytic>
	<monogr>
		<title level="m">Association for Computational Linguistics (ACL) System Demonstrations</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="55" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">How translation alters sentiment</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Mohammad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Salameh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kiritchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="95" to="130" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2018 CheckThat! Lab on automatic identification and verification of political claims</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Atanasova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kyuchukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Ninth International Conference of the CLEF Association: Experimental IR Meets Multilinguality, Multimodality, and Interaction</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">P</forename><surname>Bellot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Trabelsi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mothe</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Murtagh</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Nie</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting>the Ninth International Conference of the CLEF Association: Experimental IR Meets Multilinguality, Multimodality, and Interaction<address><addrLine>Avignon, France</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018-09">September 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">A simple but tough-tobeat baseline for the fake news challenge stance detection task</title>
		<author>
			<persName><forename type="first">B</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Augenstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">P</forename><surname>Spithourakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1707.03264</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
