<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Factoring in Context for the Automatic Detection of Misrepresentation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Bruna</forename><forename type="middle">Paz</forename><surname>Schmid</surname></persName>
							<email>bruna.pazschmid@unifr.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Fribourg</orgName>
								<address>
									<addrLine>Avenue de l&apos;Europe 20</addrLine>
									<postCode>1700</postCode>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Annette</forename><surname>Hautli-Janisz</surname></persName>
							<email>annette.hautli-janisz@uni-passau.de</email>
							<affiliation key="aff1">
								<orgName type="institution">University of Passau</orgName>
								<address>
									<postCode>94030</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Steve</forename><surname>Oswald</surname></persName>
							<email>steve.oswald@unifr.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Fribourg</orgName>
								<address>
									<addrLine>Avenue de l&apos;Europe 20</addrLine>
									<postCode>1700</postCode>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Factoring in Context for the Automatic Detection of Misrepresentation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">DB439F0E7621802B3E22CB5744C2F6D1</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Misrepresentation</term>
					<term>pragmatics</term>
					<term>natural language processing</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The aim of the paper is to show how a solid theoretical pragmatic underpinning informs an automatic approach to identifying and classifying misrepresentation in social media. To that end we present a dataset that encodes misrepresentation as well as the source that is misrepresented, building on a set of pragmatically informed annotation guidelines. The performance of standard statistic classifiers for misrepresentation detection is promising. We also perform a fine-grained manual error analysis. The paper closes with a longitudinal analysis of misrepresentation in our dataset and shows that items labelled as misrepresentation increase in years that coincide with political campaigns.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>A key feature of Trump's political campaign and one-term presidency, which started in early 2017 and ended four years later in 2021, was the strategical use of social media. In October and November 2023 alone, CNN has fact-checked twelve speeches, concluding that his "fall remarks were teeming with false claims -a staggering quantity of misrepresentations, exaggerations and outright lies that made sheer wrongness a central feature of each of his addresses" <ref type="bibr" target="#b0">[1]</ref>. Through social media, misrepresentations occur faster and in a more targeted way than in traditional media outlets where political messages are usually assessed in terms of their factual content.</p><p>Identifying misrepresentations, i.e., a metarepresentation that is not similar enough to the original representation at the inferential level given a certain context, in a systematic manner is one building block for helping voters assess the political strategies and worldviews of potential future leaders. But tackling problematic content that is spread in connection to political campaigns is not simply an issue of quickly sifting through large quantities of data. What makes it especially challenging is the quality of the content. As the results of CNN's fact-checking indicate, "sheer wrongness" comes in various forms, one of which is misrepresentation -a notion that may be understood as a form of misinformation, that is, false information.</p><p>The aim of the paper is to show how a solid theoretical pragmatic underpinning informs an automatic approach to identifying and classifying misrepresentation in social media. To this end, we present a dataset that encodes misrepresentation as well as the source that is misrepresented, building on a set of pragmatically informed annotation guidelines. The performance of standard statistic classifiers for misrepresentation detection is promising. We also perform a fine-grained manual error analysis. The paper closes with a longitudinal analysis of misrepresentation in our dataset and shows that items labelled as misrepresentation increase in years that coincide with political campaigns. The paper proceeds as follows: Section 2 discusses related work in automatically identifying misrepresentation. Section 3 presents the theoretical pragmatic underpinning, followed by a description of the annotation guidelines in Section 4. Section 5 includes information regarding data preprocessing steps, classification, model performance and error analysis. Section 6 discusses longitudinal aspects of the study, while section 7 discusses its results and Section 8 its limitations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related work</head><p>Research and governmental-driven efforts alike are trying to contain the effects of misrepresentation by finding ways to automatically identify information online that is misrepresenting the original or is outright false. For instance, the European Commission has sponsored projects aimed at developing AItools, such as Monitio, a media monitoring platform that includes fact-checking <ref type="bibr" target="#b1">[2]</ref>. The fact-checking system is evidence-based in that it works by retrieving documents from an available collection of news articles which then serve as evidence for the predictions.</p><p>However, these approaches to fact-checking and misinformation tend not to differentiate between the various forms of false information. Instead, identification often occurs with the help of vocabularies and datasets labelled based on stance detection <ref type="bibr" target="#b2">[3]</ref>, truthfulness <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>, topic-matching <ref type="bibr" target="#b6">[7]</ref>, or linguistic features associated with fake news <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>.</p><p>Misrepresentation itself has rarely been a topic of research in NLP. One exception is Michael Yeomans' study about partisan misrepresentation of political opponents through straw man arguments <ref type="bibr" target="#b9">[10]</ref>. In an experimental setting, participants were tasked with articulating their own together with their opponents' positions. They were instructed to write down open-ended responses about the Affordable Care Act. Responses were then labelled depending on whether they were the participants' genuine or imitated positions. At the computational level, texts were scored with sentiment analysis and a lexicon of morally charged words. This was followed by a machine learning model -a logistic LASSO regression -trained to distinguish between texts from opponents and supporters. However, it is difficult to conclude that the study was about the straw man argument specifically since the theoretical underpinning remains unspecified in the paper. As such, the scope of the study appears to have been limited to the analysis of partisan incendiary language accompanying misrepresentation.</p><p>In the next section, we will elaborate on our theoretical underpinning. Our approach being theorydriven is what differentiates it from others: Every step is guided by pragmatic theory because, next to identifying misrepresentation, the aim of our study was also to understand the phenomenon from a linguistic and political perspective.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Pragmatic theory and misrepresentation</head><p>The core theoretical underpinning of the present paper is an observation from pragmatics, namely that people tend to show regularity in their language use due to the social aspects of communication <ref type="bibr">[11, pp. 4-6]</ref>, for instance if people intend to misrepresent or discredit the original. This means that certain patterns in language depend on the context they are embedded in. Therefore, by defining and describing the relevant contexts, we can link patterns of language use to certain pragmatic phenomena such as misrepresentation.</p><p>In this study, we build on theories from political discourse analysis <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>, pragmatics <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref> and philosophy of language <ref type="bibr" target="#b15">[16]</ref> in defining misrepresentation as a metarepresentation that is not similar enough to the original representation at the inferential level given a certain context. Example (1) shows a tweet posted by Donald J. Trump on 27 July 2017 in which he claims that the New York Times (belonging to the left-wing spectrum) asserts that 'Fox and Friends' (right-wing spectrum) is the most powerful TV show in America.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>11-18</head><p>(1) Misrepresentation Wow, the Failing @nytimes said about @foxandfriends "....the most powerful T.V. show in America. " (ID: 890524387773997056)</p><p>The original text from an opinion piece in the NYT from 19 July 2017 is shown in <ref type="bibr" target="#b1">(2)</ref> and includes some of the text that precedes and follows the quotation:</p><p>( Based on the theoretical pragmatic underpinning of this paper, a mispresentation M in the political context needs to meet the following criteria C: C1: M is a metarepresentation in terms of intentionality.</p><p>C2: There are perceivable structural or componential differences between the original and its metarepresentation.</p><p>C3: There is noticeable change in the metarepresentation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C4:</head><p>The difference between the original representation and the metarepresentation results in a difference in comprehension.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C5:</head><p>The difference in comprehension is politically relevant.</p><p>Based on C1, Example ( <ref type="formula">1</ref>) is a metarepresentation, because it contains representative content discernible on a verbatim quotation, on the reported speech verb "said," and on the use of "[w]ow" to express a psychological state through the positive evaluation of the content of the verbatim quotation. The tweet is also significantly shorter than the overall article, satisfying C2.</p><p>Regarding C3, the quotation is isolated from the article and as a result, there is an emphasis on the content of the quotation which is evaluated positively with "[w]ow." Criteria C4 is met for a variety of reasons: For one, the quotation was taken from an opinion article and therefore represents an individual's opinion (and not necessarily NYT's). Secondly, Fox &amp; Friends being described as "the most powerful TV show in America" is surprising and concerning given the adverb "[s]uddenly", which marks unexpectedness. Removing the beginning of the sentence thus changes the overall sentiment in the tweet to a positive one that is absent in the original representation. Thirdly, the author of the original does not appear to believe that the show deserves its new status, since the latter is said to result from "no other reason than its No. 1 fan", which undermines the inherent quality of the show. Finally, from the first to the second sentence in the original, Fox &amp; Friends develops from "conservative couch gab" to "the most powerful TV show in America", i.e., the author implies that the world is turned upside down as a result of the former president's relationship with Fox &amp; Friends.</p><p>In terms of C5, a politically relevant difference emerges between the original and the misrepresentation: The New York Times' original article achieves relevance by being identified as a critical piece which deems the show's new importance to be undeserved and perhaps even dangerous, considering that it is the result of the influence of the president of the United States. From this point of view, it may be a warning against the manipulation of the media for political purposes given the media's role as a check on governmental power and democracy. This criticism is especially strong since it originates in the media itself. Strikingly, Trump's metarepresentational rendition achieves political relevance by concluding the opposite, with important contextual implications: Even the New York Times, which is biased and left-wing and is known to be critical of Trump, recognizes how important his favorite show, in which he often participates, is. Additionally, given Trump's relationship with the show, presenting it in a positive light may be an attempt to promote himself as well. Thus, it would represent an instance of positive self-presentation. Whereas the original representation is likely an attempt to revise available assumptions and thus change the current state of the world where Fox &amp; Friends is portrayed as undeservedly powerful, the metarepresentation is likely to function as an attempt to strengthen those same assumptions and thus to protect the current state of the world.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">The dataset</head><p>The basis of the investigation is the Trump Twitter Archive, a database that contains most tweets posted from Trump's personal account, @realDonaldTrump, between 2009 and 2021. The site was launched in 2016 and includes 56,571 tweets <ref type="bibr" target="#b17">[18]</ref>. The maximum character count of the tweets ranges between 140 and 280. In the following, we discuss the steps taken to prune the dataset in order to be able to model pragmatic theory and misrepresentation in a meaningful way. The dataset and classification code are available at https://github.com/runastef/auto-ident-trump-misrep.git.</p><p>Filtering Two filters are applied consecutively with the aim of increasing the likelihood that the resulting array contains misrepresentation. The first filter extracted tweets containing quotation marks, which usually signal the presence of representative content for instance in the form of reported speech. Therefore, the presence of quotation marks is more likely to be used to comment on an original representation. We exclude tweets predating Trump's presidential campaign announcement from the selection as well as retweets. The same was attempted for quoted replies by excluding tweets containing the handle @realDonaldTrump. The intention behind excluding retweets and quoted replies was to limit the conversational context of the tweets to the relevant original representations by reducing the amount of representative content. This reduces the contextual complexity of the tweets so as to strengthen the relationship between the utterances and the pragmatic phenomenon of misrepresentation. Eventually, the filter excluded the expressions 'Nobody', 'establishment', 'Washington', 'elite', and 'Congress'. The resulting pre-annotation dataset, which combines both selections, contained a total of 1,737 tweets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Annotation study</head><p>The annotation of the selected tweets was done by two annotators after instruction, one of them being a co-author of the paper. The annotation guidelines reflect the criteria for misrepresentation discussed in Section 3 in that they are the deciding factors when a tweet is judged as being a misrepresentation of an original. The decision is binary, i.e., 'misrepresentation' versus 'not-misrepresentation'. Inter-annotator agreement with Cohen's Kappa was 0.765 over the whole dataset, which signals substantive agreement. To increase the quality of the dataset only tweets that both annotators agreed on were included. The resulting dataset has 214 items, 107 are labelled as being an instance of misrepresentation and 107 not being considered misrepresentations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Predicting misrepresentation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Preprocessing</head><p>In preparation for the application of the text classification algorithms, we normalize, remove noise and anticipate and prevent issues connected to expressions such as URLs. Stop words were removed. Tokenization was done with TweetTokenizer from the NLTK library, which takes into account the specific linguistic structures prevalent in social media.</p><p>The list of stop words is updated to reflect Trump's language use. Since Trump's language use lacks complexity, removing frequent words may result in the removal of a significant amount of meaning because Trump's vocabulary contains many elements that would normally count as stop words. As a result, removing such words could influence the classification in a negative way as important patterns linked to his language use might be lost. This might cause an issue with the Naïve Bayes classifier, for example, since it is a probabilistic model that bases its decisions on the frequency with which the different tokens are present in a certain label during the training phase. Thus, to avoid losing potentially important information, most verbs and conjunctions are kept while pronouns were removed from the list of stop words.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Classification</head><p>In the next step, the TfidfVectorizer (Term-Frequency Inverse Document Frequency Vectorizer) from scikit-learn was used for vectorization, i.e., a bag of words representation of all tweets was created, containing the tf-idf values for each word across all tweets. Normally, this vectorizer does its own tokenization, i.e., a library-internal module splits the running text into tokens. For the purpose of this paper, we overwrite this module since tweets have a unique format that can be challenging for tokenization. Since scikit-learn still requires tokenization for internal reasons, we follow the method introduced in David Batista's blog and pass a dummy tokenizer and preprocessor that returns the same input without changing it in scikit-learn <ref type="bibr" target="#b18">[19]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Results</head><p>The text classification algorithms employed in this study are: Naïve Bayes, Support Vector Classifier (with a linear kernel) and Random Forest, all imported from the scikit-learn library. All classifiers were left on their standard configurations for learning purposes. On average, all three classifiers performed at around 70% based on mean accuracy (Naïve Bayes: 0.71, SVC: 0.73, Random Forest: 0.68). The accuracy scores ranged between 71% and 72% for Naïve Bayes, between 72% and 74% for SVC, and between 67% and 69% for Random Forest based on a 95% confidence interval. The scores are promising considering that a larger sample size may well improve the performance of the classifiers given that the content of the training data is expected to be more balanced with a larger sample size. The results are promising even when compared with related work: Miranda et al. 's <ref type="bibr" target="#b1">[2]</ref> evidence-based automated fact checking platform presents predictions maintained to be correct 58% of the time, and Pérez-Rosas et al.'s <ref type="bibr" target="#b8">[9]</ref> fake news detector reportedly presents accuracies between 50% and 76% depending on the domain associated with the dataset that is used. The results also suggest that the Support Vector Machine classifier performed slightly better than the other two.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4.">Error analysis</head><p>The performances of Naïve Bayes and SVC are slightly reduced in testing sets containing larger tweets. That is, the performance seems to decrease the higher the number of tokens in the testing set is. This could be due to an imbalance between the training set and the testing set. Random Forest was probably less affected by this because it relies on the decision of multiple classifiers that reach their individual decisions based on a significantly smaller sample size than SVC and Naïve Bayes during the same run.</p><p>Incorrect predictions often arise when the language in the tweet is uncharacteristically complex, for instance when a supporter's tweet is copied and posted through Trump's account and contains language that is more complex than Trump's typical writing style. Here is such an example:</p><p>(3)</p><p>Copied and posted tweet "@racheljoycowley: I'm done with Macy's. Apparently, they follow the trend of trying to force legislation on every American freedom. Done!" (ID: 616646192609558528)</p><p>In an arbitrary run, this tweet was incorrectly predicted as not-misrepresentation by all three classifiers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Tracking misrepresentation over time</head><p>Having annotated the tweets to be either misrepresentation or not, counting them over the years on the x-axis allows for a longitudinal view of the tweeting behavior of Donald J. Trump. Figure <ref type="figure" target="#fig_0">1</ref> hints at a possible correlation between the number of tweets containing misrepresentations that Trump posted and the years in which he was active in presidential election campaigns. The overall trend towards less tweets was already evident in the pre-annotation dataset. Thus, it is not surprising. What is interesting is the way in which the trend appears to change abruptly in 2018 and then again in 2019. Both categories experience an increase between 2018 and 2019. However, whereas the instances of not-misrepresentation begin to decrease from 2019 on, the instances of misrepresentation rise sharply.</p><p>It is also worth pointing out that given the connection to campaign years, the graph may be interpreted in terms of the reliability of the created dataset since misrepresentation is probably more likely to occur during political campaigns due to the nature of politics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Discussion and summary</head><p>Future work could include a general analysis of Trump's language use, which might help to improve the preprocessing. This could take the form of a linguistic features engineering study to determine his writing style. The findings could then be used to create a stop words list that is better able to reflect Trump's language use although, perhaps, a more generalized approach based on country and political affiliation may be more helpful for the study of misrepresentation itself because it would be easier to generalize. It may also prove to be more practical in terms of implementation.</p><p>The process of evaluating the performances of the classifiers could be simplified and improved with an explainer. The LIME (Local Interpretable Model-Agnostic Explanations) Text Explainer was difficult to implement even after replacing the LinearSVC classifier with an SVC classifier with a linear kernel. Initially, the LinearSVC classifier was used, but it was changed into an SVC classifier with a linear kernel. It should produce similar results seeing as LinearSVC is an implementation of SVC. The change was necessary because LinearSVC does not support the function predict_proba, which calculates the probabilities for each class prediction. It would have been difficult to evaluate the performance of the LinearSVC classifier without predict_proba. LIME was chosen because, in theory, it should work well with all three classifiers as long as the models are able to "predict the probabilities of the categories." According to Albrecht et al., LIME "works locally by taking a look at each prediction separately. This is achieved by modifying the input vector to find the local components that the predictions are sensitive to" <ref type="bibr">[20, p. 195</ref>]. Then, "[f]rom the behavior in the vicinity of the vector, it will draw conclusions about which components are more or less important" and "visualize the contributions and explain the decision mechanism of the algorithm for individual documents" <ref type="bibr">[20, p. 195</ref>]. However, LIME's implementation and interpretation was challenging. As such, in the end, it was not taken into consideration in the error analysis. And yet, evaluating the performance of the classifiers would probably have been significantly more straightforward with such an explainer. In its absence, the process is considerably slowed down. Improving explainability would have practical implications for future research, for instance, although we favor a theory-driven approach, together with the misrepresentation dataset the findings could be used to improve or expand available vocabularies employed in current fact-checking systems.</p><p>To summarize, this study contributes to the research of pragmatically relevant phenomena with computational linguistic methods by discussing how to account for various aspects of the context at different stages of the research process. Specifically, efforts were made to retain contextual information related to the social and political contexts. To this end, a framework was developed for the pragmatic analysis of political misrepresentation with computational methods. Based on the framework, annotation guidelines were written to enable the creation of a misrepresentation dataset, which was then employed to train supervised machine learning algorithms used for text classification. The three algorithms performed at around 70% with SVC performing slightly better than the other two algorithms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Limitations</head><p>The findings of this study may be limited by the small size of the dataset, the data selection process, data source, and the format of the text data.</p><p>The study relied on tweets posted from Trump's Twitter account. Consequently, the methods applied in this study might yield different results on the discourses of other individuals especially if one considers Trump's unique language use and political affiliation. Widening the scope of the study to include political discourse from a larger number of politicians is likely to lead to better insight into political misrepresentation. To this end, our study will hopefully provide a basis for further research into a topic that has not received a lot of attention so far.</p><p>The relative novelty of pragmatic research into misrepresentation with computational linguistic methods also explains the chosen format. The smaller number of tokens present in tweets were expected to facilitate computation given the theoretical underpinning. Although the small format may limit the study's generalizability, it facilitates the qualitative analysis of the results which will help us to widen the scope of the analysis in future studies on this topic.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Number of Misrepresentations over time</figDesc><graphic coords="6,184.82,65.61,225.63,188.24" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Trump&apos;s avalanche of dishonesty: Fact-checking 102 of his false claims from this fall</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<ptr target="https://edition.cnn.com/2023/12/01/politics/trump-dishonesty-avalanche-102-fall-false-claims/index.html" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Automated fact checking in the news room</title>
		<author>
			<persName><forename type="first">S</forename><surname>Miranda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nogueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mendes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vlachos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Secker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Garrett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitchel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Marinho</surname></persName>
		</author>
		<idno type="DOI">10.1145/3308558.3314135</idno>
	</analytic>
	<monogr>
		<title level="m">The World Wide Web Conference</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>White</surname></persName>
		</editor>
		<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3579" to="3583" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Emergent: a novel data-set for stance classification</title>
		<author>
			<persName><forename type="first">W</forename><surname>Ferreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vlachos</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N16-1138</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Knight</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Nenkova</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">O</forename><surname>Rambow</surname></persName>
		</editor>
		<meeting>the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics<address><addrLine>San Diego, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1163" to="1168" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Truth of varying shades: Analyzing language in fake news and political fact-checking</title>
		<author>
			<persName><forename type="first">H</forename><surname>Rashkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Jang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Volkova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D17-1317</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Palmer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Hwa</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</editor>
		<meeting>the 2017 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics<address><addrLine>Copenhagen, Denmark</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="2931" to="2937" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">liar, liar pants on fire&quot;: A new benchmark dataset for fake news detection</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P17-2067</idno>
		<ptr target="https://aclanthology.org/P17-2067.doi:10.18653/v1/P17-2067" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics: Short Papers, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Barzilay</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M.-Y</forename><surname>Kan</surname></persName>
		</editor>
		<meeting>the 55th Annual Meeting of the Association for Computational Linguistics: Short Papers, Association for Computational Linguistics<address><addrLine>Vancouver, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="422" to="426" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A language-based approach to fake news detection through interpretable features and brnn</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Qiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wiechmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kerz</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.rdsm-1.2" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd International Workshop on Rumours and Deception in Social Media (RDSM), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Aker</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</editor>
		<meeting>the 3rd International Workshop on Rumours and Deception in Social Media (RDSM), Association for Computational Linguistics<address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="14" to="31" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Loco: The 88-million-word language of conspiracy corpus</title>
		<author>
			<persName><forename type="first">A</forename><surname>Miani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hills</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bangerter</surname></persName>
		</author>
		<idno type="DOI">10.3758/s13428-021-01698-z</idno>
	</analytic>
	<monogr>
		<title level="j">Behavior research methods</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="1794" to="1817" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Fake news detection for the russian language</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kuzmin</surname></persName>
		</author>
		<author>
			<persName><surname>Larionov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Daniil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Pisarevskaya</surname></persName>
		</author>
		<author>
			<persName><surname>Smirnov</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.rdsm-1.5/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd International Workshop on Rumours and Deception in Social Media (RDSM), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Aker</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</editor>
		<meeting>the 3rd International Workshop on Rumours and Deception in Social Media (RDSM), Association for Computational Linguistics<address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="45" to="57" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Automatic detection of fake news</title>
		<author>
			<persName><forename type="first">V</forename><surname>Pérez-Rosas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kleinberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lefevre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mihalcea</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/C18-1287" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th International Conference on Computational Linguistics, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Bender</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Derczynski</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Isabelle</surname></persName>
		</editor>
		<meeting>the 27th International Conference on Computational Linguistics, Association for Computational Linguistics<address><addrLine>Santa Fe, NM, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="3391" to="3401" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The straw man effect: Partisan misrepresentation in natural language</title>
		<author>
			<persName><forename type="first">M</forename><surname>Yeomans</surname></persName>
		</author>
		<idno type="DOI">10.1177/13684302211014582</idno>
	</analytic>
	<monogr>
		<title level="j">Group Processes &amp; Intergroup Relations</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="1905" to="1924" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Yule</surname></persName>
		</author>
		<title level="m">Pragmatics, Oxford introductions to language study</title>
				<meeting><address><addrLine>Oxford, UK</addrLine></address></meeting>
		<imprint>
			<publisher>Oxford University Press</publisher>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Ideology and discourse analysis</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">A</forename><surname>Van Dijk</surname></persName>
		</author>
		<idno type="DOI">10.1080/13569310600687908</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Political Ideologies</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="115" to="140" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Political discourse</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wilson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Handbook of Discourse Analysis, Blackwell Handbooks in Linguistics</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Tannen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">E</forename><surname>Hamilton</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schiffrin</surname></persName>
		</editor>
		<meeting><address><addrLine>Malden and Oxford</addrLine></address></meeting>
		<imprint>
			<publisher>Wiley Blackwell</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="775" to="794" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Metarepresentation: A Relevance-Theory Approach</title>
		<author>
			<persName><forename type="first">E.-J</forename><surname>Noh</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>John Benjamins Publishing Company</publisher>
			<biblScope unit="volume">69</biblScope>
			<pubPlace>Amsterdam</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Relevance theory</title>
		<author>
			<persName><forename type="first">D</forename><surname>Wilson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sperber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Handbook of Pragmatics</title>
				<editor>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Horn</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Ward</surname></persName>
		</editor>
		<meeting><address><addrLine>Malden, MA</addrLine></address></meeting>
		<imprint>
			<publisher>Wiley</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="607" to="632" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Intentionality: An Essay in the Philosophy of Mind</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Searle</surname></persName>
		</author>
		<idno type="DOI">10.1017/CBO9781139173452</idno>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Cambridge University Press</publisher>
			<pubPlace>Cambridge, UK</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Watching &apos;fox &amp; friends,&apos; trump sees a two-way mirror</title>
		<author>
			<persName><forename type="first">J</forename><surname>Poniewozik</surname></persName>
		</author>
		<ptr target="https://www.nytimes.com/2017/07/19/arts/television/donald-trump-fox-friends.html" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Brown</surname></persName>
		</author>
		<ptr target="https://www.thetrumparchive.com/" />
		<title level="m">Trump twitter archive</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Applying scikit-learn tfidfvectorizer on tokenized text</title>
		<author>
			<persName><forename type="first">D</forename><surname>Batista</surname></persName>
		</author>
		<ptr target="https://www.davidsbatista.net/blog/2018/02/28/TfidfVectorizer/" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Blueprints for Text Analytics Using Python: Machine Learning-Based Solutions for Common Real World (NLP) Applications</title>
		<author>
			<persName><forename type="first">J</forename><surname>Albrecht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ramachandran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Winkler</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<publisher>O&apos;Reilly Media, Inc</publisher>
			<pubPlace>Sebastopol, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
