=Paper=
{{Paper
|id=Vol-3782/paper2
|storemode=property
|title=Factoring in Context for the Automatic Detection of Misrepresentation
|pdfUrl=https://ceur-ws.org/Vol-3782/paper2short.pdf
|volume=Vol-3782
|authors=Bruna Paz Schmid,Annette Hautli-Janisz,Steve Oswald
|dblpUrl=https://dblp.org/rec/conf/codai2/SchmidHO24
}}
==Factoring in Context for the Automatic Detection of Misrepresentation==
Factoring in Context for the Automatic Detection of
Misrepresentation
Bruna Paz Schmid1,* , Annette Hautli-Janisz2 and Steve Oswald1
1
University of Fribourg, Avenue de l’Europe 20, 1700 Fribourg, Switzerland
2
University of Passau, 94030 Passau, Germany
Abstract
The aim of the paper is to show how a solid theoretical pragmatic underpinning informs an automatic approach
to identifying and classifying misrepresentation in social media. To that end we present a dataset that encodes
misrepresentation as well as the source that is misrepresented, building on a set of pragmatically informed
annotation guidelines. The performance of standard statistic classifiers for misrepresentation detection is
promising. We also perform a fine-grained manual error analysis. The paper closes with a longitudinal analysis
of misrepresentation in our dataset and shows that items labelled as misrepresentation increase in years that
coincide with political campaigns.
Keywords
Misrepresentation, pragmatics, natural language processing
1. Introduction
A key feature of Trump’s political campaign and one-term presidency, which started in early 2017 and
ended four years later in 2021, was the strategical use of social media. In October and November 2023
alone, CNN has fact-checked twelve speeches, concluding that his “fall remarks were teeming with false
claims - a staggering quantity of misrepresentations, exaggerations and outright lies that made sheer
wrongness a central feature of each of his addresses” [1]. Through social media, misrepresentations
occur faster and in a more targeted way than in traditional media outlets where political messages are
usually assessed in terms of their factual content.
Identifying misrepresentations, i.e., a metarepresentation that is not similar enough to the original
representation at the inferential level given a certain context, in a systematic manner is one building
block for helping voters assess the political strategies and worldviews of potential future leaders. But
tackling problematic content that is spread in connection to political campaigns is not simply an issue
of quickly sifting through large quantities of data. What makes it especially challenging is the quality of
the content. As the results of CNN’s fact-checking indicate, “sheer wrongness” comes in various forms,
one of which is misrepresentation – a notion that may be understood as a form of misinformation, that
is, false information.
The aim of the paper is to show how a solid theoretical pragmatic underpinning informs an automatic
approach to identifying and classifying misrepresentation in social media. To this end, we present a
dataset that encodes misrepresentation as well as the source that is misrepresented, building on a set
of pragmatically informed annotation guidelines. The performance of standard statistic classifiers for
misrepresentation detection is promising. We also perform a fine-grained manual error analysis. The
paper closes with a longitudinal analysis of misrepresentation in our dataset and shows that items
labelled as misrepresentation increase in years that coincide with political campaigns.
Proceedings of the 1st Workshop on COuntering Disinformation with Artificial Intelligence (CODAI), co-located with the 27th
European Conference on Artificial Intelligence (ECAI), pages 11–18, October 20, 2024, Santiago de Compostela, Spain
*
Corresponding author.
$ bruna.pazschmid@unifr.ch (B. Paz Schmid); annette.hautli-janisz@uni-passau.de (A. Hautli-Janisz);
steve.oswald@unifr.ch (S. Oswald)
https://www.fim.uni-passau.de/en/cornlp (A. Hautli-Janisz); https://www.steveoswald.ch (S. Oswald)
0009-0009-7734-0699 (B. Paz Schmid); 0000-0002-5901-9633 (A. Hautli-Janisz); 0000-0002-5946-1691 (S. Oswald)
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
11
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
The paper proceeds as follows: Section 2 discusses related work in automatically identifying misrep-
resentation. Section 3 presents the theoretical pragmatic underpinning, followed by a description of the
annotation guidelines in Section 4. Section 5 includes information regarding data preprocessing steps,
classification, model performance and error analysis. Section 6 discusses longitudinal aspects of the
study, while section 7 discusses its results and Section 8 its limitations.
2. Related work
Research and governmental-driven efforts alike are trying to contain the effects of misrepresentation
by finding ways to automatically identify information online that is misrepresenting the original or is
outright false. For instance, the European Commission has sponsored projects aimed at developing AI-
tools, such as Monitio, a media monitoring platform that includes fact-checking [2]. The fact-checking
system is evidence-based in that it works by retrieving documents from an available collection of news
articles which then serve as evidence for the predictions.
However, these approaches to fact-checking and misinformation tend not to differentiate between the
various forms of false information. Instead, identification often occurs with the help of vocabularies and
datasets labelled based on stance detection [3], truthfulness [4, 5, 6], topic-matching [7], or linguistic
features associated with fake news [8, 9].
Misrepresentation itself has rarely been a topic of research in NLP. One exception is Michael Yeomans’
study about partisan misrepresentation of political opponents through straw man arguments [10]. In an
experimental setting, participants were tasked with articulating their own together with their opponents’
positions. They were instructed to write down open-ended responses about the Affordable Care Act.
Responses were then labelled depending on whether they were the participants’ genuine or imitated
positions. At the computational level, texts were scored with sentiment analysis and a lexicon of morally
charged words. This was followed by a machine learning model – a logistic LASSO regression - trained
to distinguish between texts from opponents and supporters. However, it is difficult to conclude that
the study was about the straw man argument specifically since the theoretical underpinning remains
unspecified in the paper. As such, the scope of the study appears to have been limited to the analysis of
partisan incendiary language accompanying misrepresentation.
In the next section, we will elaborate on our theoretical underpinning. Our approach being theory-
driven is what differentiates it from others: Every step is guided by pragmatic theory because, next to
identifying misrepresentation, the aim of our study was also to understand the phenomenon from a
linguistic and political perspective.
3. Pragmatic theory and misrepresentation
The core theoretical underpinning of the present paper is an observation from pragmatics, namely
that people tend to show regularity in their language use due to the social aspects of communication
[11, pp. 4-6], for instance if people intend to misrepresent or discredit the original. This means that
certain patterns in language depend on the context they are embedded in. Therefore, by defining and
describing the relevant contexts, we can link patterns of language use to certain pragmatic phenomena
such as misrepresentation.
In this study, we build on theories from political discourse analysis [12, 13], pragmatics [14, 15]
and philosophy of language [16] in defining misrepresentation as a metarepresentation that is not
similar enough to the original representation at the inferential level given a certain context. Example
(1) shows a tweet posted by Donald J. Trump on 27 July 2017 in which he claims that the New York
Times (belonging to the left-wing spectrum) asserts that ‘Fox and Friends’ (right-wing spectrum) is the
most powerful TV show in America.
12
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
(1) Misrepresentation
Wow, the Failing @nytimes said about @foxandfriends "....the most powerful T.V. show in America."
(ID: 890524387773997056)
The original text from an opinion piece in the NYT from 19 July 2017 is shown in (2) and includes some
of the text that precedes and follows the quotation:
(2) Original
For years, it was a nontaxing mix of news, lifestyle and conservative couch gab, a warm-up before
Fox’s day of politics and commentary. Suddenly, for no other reason than its No. 1 fan, it is
the most powerful TV show in America. (It’s also easily the most-watched cable news morning
show, averaging 1.6 million viewers in the year’s second quarter, following a post-Trump ratings
boost.) [...] [17]
Based on the theoretical pragmatic underpinning of this paper, a mispresentation M in the political
context needs to meet the following criteria C:
C1: M is a metarepresentation in terms of intentionality.
C2: There are perceivable structural or componential differences between the original and its metarep-
resentation.
C3: There is noticeable change in the metarepresentation.
C4: The difference between the original representation and the metarepresentation results in a differ-
ence in comprehension.
C5: The difference in comprehension is politically relevant.
Based on C1, Example (1) is a metarepresentation, because it contains representative content dis-
cernible on a verbatim quotation, on the reported speech verb “said,” and on the use of “[w]ow” to
express a psychological state through the positive evaluation of the content of the verbatim quotation.
The tweet is also significantly shorter than the overall article, satisfying C2.
Regarding C3, the quotation is isolated from the article and as a result, there is an emphasis on the
content of the quotation which is evaluated positively with “[w]ow.” Criteria C4 is met for a variety
of reasons: For one, the quotation was taken from an opinion article and therefore represents an
individual’s opinion (and not necessarily NYT’s). Secondly, Fox & Friends being described as “the most
powerful TV show in America” is surprising and concerning given the adverb “[s]uddenly”, which
marks unexpectedness. Removing the beginning of the sentence thus changes the overall sentiment
in the tweet to a positive one that is absent in the original representation. Thirdly, the author of the
original does not appear to believe that the show deserves its new status, since the latter is said to result
from “no other reason than its No. 1 fan”, which undermines the inherent quality of the show. Finally,
from the first to the second sentence in the original, Fox & Friends develops from “conservative couch
gab” to “the most powerful TV show in America”, i.e., the author implies that the world is turned upside
down as a result of the former president’s relationship with Fox & Friends.
In terms of C5, a politically relevant difference emerges between the original and the misrepresen-
tation: The New York Times’ original article achieves relevance by being identified as a critical piece
which deems the show’s new importance to be undeserved and perhaps even dangerous, considering
that it is the result of the influence of the president of the United States. From this point of view,
it may be a warning against the manipulation of the media for political purposes given the media’s
role as a check on governmental power and democracy. This criticism is especially strong since it
originates in the media itself. Strikingly, Trump’s metarepresentational rendition achieves political
relevance by concluding the opposite, with important contextual implications: Even the New York
Times, which is biased and left-wing and is known to be critical of Trump, recognizes how important
his favorite show, in which he often participates, is. Additionally, given Trump’s relationship with the
13
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
show, presenting it in a positive light may be an attempt to promote himself as well. Thus, it would
represent an instance of positive self-presentation. Whereas the original representation is likely an
attempt to revise available assumptions and thus change the current state of the world where Fox
& Friends is portrayed as undeservedly powerful, the metarepresentation is likely to function as an
attempt to strengthen those same assumptions and thus to protect the current state of the world.
4. The dataset
The basis of the investigation is the Trump Twitter Archive, a database that contains most tweets posted
from Trump’s personal account, @realDonaldTrump, between 2009 and 2021. The site was launched in
2016 and includes 56,571 tweets [18]. The maximum character count of the tweets ranges between 140
and 280. In the following, we discuss the steps taken to prune the dataset in order to be able to model
pragmatic theory and misrepresentation in a meaningful way. The dataset and classification code are
available at https://github.com/runastef/auto-ident-trump-misrep.git.
Filtering Two filters are applied consecutively with the aim of increasing the likelihood that the
resulting array contains misrepresentation. The first filter extracted tweets containing quotation marks,
which usually signal the presence of representative content for instance in the form of reported speech.
Therefore, the presence of quotation marks is more likely to be used to comment on an original
representation. We exclude tweets predating Trump’s presidential campaign announcement from the
selection as well as retweets. The same was attempted for quoted replies by excluding tweets containing
the handle @realDonaldTrump. The intention behind excluding retweets and quoted replies was to limit
the conversational context of the tweets to the relevant original representations by reducing the amount
of representative content. This reduces the contextual complexity of the tweets so as to strengthen the
relationship between the utterances and the pragmatic phenomenon of misrepresentation. Eventually,
the filter excluded the expressions ‘Nobody’, ‘establishment’, ‘Washington’, ‘elite’, and ‘Congress’. The
resulting pre-annotation dataset, which combines both selections, contained a total of 1,737 tweets.
Annotation study The annotation of the selected tweets was done by two annotators after instruc-
tion, one of them being a co-author of the paper. The annotation guidelines reflect the criteria for
misrepresentation discussed in Section 3 in that they are the deciding factors when a tweet is judged
as being a misrepresentation of an original. The decision is binary, i.e., ’misrepresentation’ versus
’not-misrepresentation’. Inter-annotator agreement with Cohen’s Kappa was 0.765 over the whole
dataset, which signals substantive agreement. To increase the quality of the dataset only tweets that
both annotators agreed on were included. The resulting dataset has 214 items, 107 are labelled as being
an instance of misrepresentation and 107 not being considered misrepresentations.
5. Predicting misrepresentation
5.1. Preprocessing
In preparation for the application of the text classification algorithms, we normalize, remove noise
and anticipate and prevent issues connected to expressions such as URLs. Stop words were removed.
Tokenization was done with TweetTokenizer from the NLTK library, which takes into account the
specific linguistic structures prevalent in social media.
The list of stop words is updated to reflect Trump’s language use. Since Trump’s language use lacks
complexity, removing frequent words may result in the removal of a significant amount of meaning
because Trump’s vocabulary contains many elements that would normally count as stop words. As a
result, removing such words could influence the classification in a negative way as important patterns
linked to his language use might be lost. This might cause an issue with the Naïve Bayes classifier,
for example, since it is a probabilistic model that bases its decisions on the frequency with which the
14
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
different tokens are present in a certain label during the training phase. Thus, to avoid losing potentially
important information, most verbs and conjunctions are kept while pronouns were removed from the
list of stop words.
5.2. Classification
In the next step, the TfidfVectorizer (Term-Frequency Inverse Document Frequency Vectorizer) from
scikit-learn was used for vectorization, i.e., a bag of words representation of all tweets was created,
containing the tf-idf values for each word across all tweets. Normally, this vectorizer does its own
tokenization, i.e., a library-internal module splits the running text into tokens. For the purpose of
this paper, we overwrite this module since tweets have a unique format that can be challenging for
tokenization. Since scikit-learn still requires tokenization for internal reasons, we follow the method
introduced in David Batista’s blog and pass a dummy tokenizer and preprocessor that returns the same
input without changing it in scikit-learn [19].
5.3. Results
The text classification algorithms employed in this study are: Naïve Bayes, Support Vector Classifier
(with a linear kernel) and Random Forest, all imported from the scikit-learn library. All classifiers were
left on their standard configurations for learning purposes. On average, all three classifiers performed at
around 70% based on mean accuracy (Naïve Bayes: 0.71, SVC: 0.73, Random Forest: 0.68). The accuracy
scores ranged between 71% and 72% for Naïve Bayes, between 72% and 74% for SVC, and between 67%
and 69% for Random Forest based on a 95% confidence interval. The scores are promising considering
that a larger sample size may well improve the performance of the classifiers given that the content of
the training data is expected to be more balanced with a larger sample size. The results are promising
even when compared with related work: Miranda et al.’s [2] evidence-based automated fact checking
platform presents predictions maintained to be correct 58% of the time, and Pérez-Rosas et al.’s [9]
fake news detector reportedly presents accuracies between 50% and 76% depending on the domain
associated with the dataset that is used. The results also suggest that the Support Vector Machine
classifier performed slightly better than the other two.
5.4. Error analysis
The performances of Naïve Bayes and SVC are slightly reduced in testing sets containing larger tweets.
That is, the performance seems to decrease the higher the number of tokens in the testing set is. This
could be due to an imbalance between the training set and the testing set. Random Forest was probably
less affected by this because it relies on the decision of multiple classifiers that reach their individual
decisions based on a significantly smaller sample size than SVC and Naïve Bayes during the same run.
Incorrect predictions often arise when the language in the tweet is uncharacteristically complex,
for instance when a supporter’s tweet is copied and posted through Trump’s account and contains
language that is more complex than Trump’s typical writing style. Here is such an example:
(3) Copied and posted tweet
"@racheljoycowley: I’m done with Macy’s. Apparently, they follow the trend of trying to force
legislation on every American freedom. Done!" (ID: 616646192609558528)
In an arbitrary run, this tweet was incorrectly predicted as not-misrepresentation by all three classifiers.
6. Tracking misrepresentation over time
Having annotated the tweets to be either misrepresentation or not, counting them over the years on
the x-axis allows for a longitudinal view of the tweeting behavior of Donald J. Trump.
15
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
Figure 1: Number of Misrepresentations over time
Figure 1 hints at a possible correlation between the number of tweets containing misrepresentations
that Trump posted and the years in which he was active in presidential election campaigns. The overall
trend towards less tweets was already evident in the pre-annotation dataset. Thus, it is not surprising.
What is interesting is the way in which the trend appears to change abruptly in 2018 and then again in
2019. Both categories experience an increase between 2018 and 2019. However, whereas the instances of
not-misrepresentation begin to decrease from 2019 on, the instances of misrepresentation rise sharply.
It is also worth pointing out that given the connection to campaign years, the graph may be interpreted
in terms of the reliability of the created dataset since misrepresentation is probably more likely to occur
during political campaigns due to the nature of politics.
7. Discussion and summary
Future work could include a general analysis of Trump’s language use, which might help to improve
the preprocessing. This could take the form of a linguistic features engineering study to determine his
writing style. The findings could then be used to create a stop words list that is better able to reflect
Trump’s language use although, perhaps, a more generalized approach based on country and political
affiliation may be more helpful for the study of misrepresentation itself because it would be easier to
generalize. It may also prove to be more practical in terms of implementation.
The process of evaluating the performances of the classifiers could be simplified and improved with
an explainer. The LIME (Local Interpretable Model-Agnostic Explanations) Text Explainer was difficult
to implement even after replacing the LinearSVC classifier with an SVC classifier with a linear kernel.
Initially, the LinearSVC classifier was used, but it was changed into an SVC classifier with a linear
kernel. It should produce similar results seeing as LinearSVC is an implementation of SVC. The change
was necessary because LinearSVC does not support the function predict_proba, which calculates the
probabilities for each class prediction. It would have been difficult to evaluate the performance of the
LinearSVC classifier without predict_proba. LIME was chosen because, in theory, it should work well
with all three classifiers as long as the models are able to “predict the probabilities of the categories.”
According to Albrecht et al., LIME “works locally by taking a look at each prediction separately. This is
achieved by modifying the input vector to find the local components that the predictions are sensitive
to” [20, p. 195]. Then, “[f]rom the behavior in the vicinity of the vector, it will draw conclusions about
which components are more or less important” and “visualize the contributions and explain the decision
mechanism of the algorithm for individual documents” [20, p. 195]. However, LIME’s implementation
and interpretation was challenging. As such, in the end, it was not taken into consideration in the error
analysis. And yet, evaluating the performance of the classifiers would probably have been significantly
16
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
more straightforward with such an explainer. In its absence, the process is considerably slowed down.
Improving explainability would have practical implications for future research, for instance, although
we favor a theory-driven approach, together with the misrepresentation dataset the findings could be
used to improve or expand available vocabularies employed in current fact-checking systems.
To summarize, this study contributes to the research of pragmatically relevant phenomena with
computational linguistic methods by discussing how to account for various aspects of the context at
different stages of the research process. Specifically, efforts were made to retain contextual information
related to the social and political contexts. To this end, a framework was developed for the pragmatic
analysis of political misrepresentation with computational methods. Based on the framework, annotation
guidelines were written to enable the creation of a misrepresentation dataset, which was then employed
to train supervised machine learning algorithms used for text classification. The three algorithms
performed at around 70% with SVC performing slightly better than the other two algorithms.
8. Limitations
The findings of this study may be limited by the small size of the dataset, the data selection process,
data source, and the format of the text data.
The study relied on tweets posted from Trump’s Twitter account. Consequently, the methods applied
in this study might yield different results on the discourses of other individuals especially if one
considers Trump’s unique language use and political affiliation. Widening the scope of the study to
include political discourse from a larger number of politicians is likely to lead to better insight into
political misrepresentation. To this end, our study will hopefully provide a basis for further research
into a topic that has not received a lot of attention so far.
The relative novelty of pragmatic research into misrepresentation with computational linguistic
methods also explains the chosen format. The smaller number of tokens present in tweets were expected
to facilitate computation given the theoretical underpinning. Although the small format may limit the
study’s generalizability, it facilitates the qualitative analysis of the results which will help us to widen
the scope of the analysis in future studies on this topic.
References
[1] D. Dale, Trump’s avalanche of dishonesty: Fact-checking 102 of his false
claims from this fall, 2023. URL: https://edition.cnn.com/2023/12/01/politics/
trump-dishonesty-avalanche-102-fall-false-claims/index.html.
[2] S. Miranda, D. Nogueira, A. Mendes, A. Vlachos, A. Secker, R. Garrett, J. Mitchel, Z. Marinho,
Automated fact checking in the news room, in: L. Liu, R. White (Eds.), The World Wide Web
Conference, ACM, New York, NY, USA, 2019, pp. 3579–3583. doi:10.1145/3308558.3314135.
[3] W. Ferreira, A. Vlachos, Emergent: a novel data-set for stance classification, in: K. Knight,
A. Nenkova, O. Rambow (Eds.), Proceedings of the 2016 Conference of the North American Chapter
of the Association for Computational Linguistics: Human Language Technologies, Association
for Computational Linguistics, San Diego, CA, USA, 2016, pp. 1163–1168. doi:10.18653/v1/
N16-1138.
[4] H. Rashkin, E. Choi, J. Y. Jang, S. Volkova, Y. Choi, Truth of varying shades: Analyzing language
in fake news and political fact-checking, in: M. Palmer, R. Hwa, S. Riedel (Eds.), Proceedings
of the 2017 Conference on Empirical Methods in Natural Language Processing, Association for
Computational Linguistics, Copenhagen, Denmark, 2017, pp. 2931–2937. doi:10.18653/v1/
D17-1317.
[5] W. Y. Wang, "liar, liar pants on fire": A new benchmark dataset for fake news detection, in: R. Barzi-
lay, M.-Y. Kan (Eds.), Proceedings of the 55th Annual Meeting of the Association for Computational
Linguistics: Short Papers, Association for Computational Linguistics, Vancouver, Canada, 2017, pp.
422–426. URL: https://aclanthology.org/P17-2067. doi:10.18653/v1/P17-2067.
17
Bruna Paz Schmid et al. CODAI Workshop Proceedings 11–18
[6] Y. Qiao, D. Wiechmann, E. Kerz, A language-based approach to fake news detection through
interpretable features and brnn, in: A. Aker, A. Zubiaga (Eds.), Proceedings of the 3rd International
Workshop on Rumours and Deception in Social Media (RDSM), Association for Computational
Linguistics, Barcelona, Spain, 2020, pp. 14–31. URL: https://aclanthology.org/2020.rdsm-1.2.
[7] A. Miani, T. Hills, A. Bangerter, Loco: The 88-million-word language of conspiracy corpus,
Behavior research methods 54 (2022) 1794–1817. doi:10.3758/s13428-021-01698-z.
[8] G. Kuzmin, Larionov, Daniil, D. Pisarevskaya, I. Smirnov, Fake news detection for the russian
language, in: A. Aker, A. Zubiaga (Eds.), Proceedings of the 3rd International Workshop on
Rumours and Deception in Social Media (RDSM), Association for Computational Linguistics,
Barcelona, Spain, 2020, pp. 45–57. URL: https://aclanthology.org/2020.rdsm-1.5/.
[9] V. Pérez-Rosas, B. Kleinberg, A. Lefevre, R. Mihalcea, Automatic detection of fake news, in: E. M.
Bender, L. Derczynski, P. Isabelle (Eds.), Proceedings of the 27th International Conference on
Computational Linguistics, Association for Computational Linguistics, Santa Fe, NM, USA, 2018,
pp. 3391–3401. URL: https://aclanthology.org/C18-1287.
[10] M. Yeomans, The straw man effect: Partisan misrepresentation in natural language, Group
Processes & Intergroup Relations 25 (2022) 1905–1924. doi:10.1177/13684302211014582.
[11] G. Yule, Pragmatics, Oxford introductions to language study, Oxford University Press, Oxford, UK,
2011.
[12] T. A. van Dijk, Ideology and discourse analysis, Journal of Political Ideologies 11 (2006) 115–140.
doi:10.1080/13569310600687908.
[13] J. Wilson, Political discourse, in: D. Tannen, H. E. Hamilton, D. Schiffrin (Eds.), The Handbook of
Discourse Analysis, Blackwell Handbooks in Linguistics, Wiley Blackwell, Malden and Oxford,
2015, pp. 775–794.
[14] E.-J. Noh, Metarepresentation: A Relevance-Theory Approach, volume 69, John Benjamins Pub-
lishing Company, Amsterdam, 2000.
[15] D. Wilson, D. Sperber, Relevance theory, in: L. R. Horn, G. Ward (Eds.), The Handbook of
Pragmatics, Wiley, Malden, MA, 2006, pp. 607–632.
[16] J. R. Searle, Intentionality: An Essay in the Philosophy of Mind, Cambridge University Press,
Cambridge, UK, 2012. doi:10.1017/CBO9781139173452.
[17] J. Poniewozik, Watching ’fox & friends,’ trump sees a two-way mirror, 2017. URL: https://www.
nytimes.com/2017/07/19/arts/television/donald-trump-fox-friends.html.
[18] B. Brown, Trump twitter archive, 2016. URL: https://www.thetrumparchive.com/.
[19] D. Batista, Applying scikit-learn tfidfvectorizer on tokenized text, 2018. URL: https://www.
davidsbatista.net/blog/2018/02/28/TfidfVectorizer/.
[20] J. Albrecht, S. Ramachandran, C. Winkler, Blueprints for Text Analytics Using Python: Machine
Learning-Based Solutions for Common Real World (NLP) Applications, O’Reilly Media, Inc.,
Sebastopol, CA, 2021.
18