=Paper=
{{Paper
|id=Vol-2882/MediaEval_20_paper_10
|storemode=property
|title=Detecting
Conspiracy Tweets Using Support Vector Machines
|pdfUrl=https://ceur-ws.org/Vol-2882/paper10.pdf
|volume=Vol-2882
|authors=Manfred
Moosleitner,Benjamin Murauer,Günther
Specht
|dblpUrl=https://dblp.org/rec/conf/mediaeval/MoosleitnerMS20
}}
==Detecting
Conspiracy Tweets Using Support Vector Machines==
Detecting Conspiracy Tweets Using Support Vector Machines
Manfred Moosleitner, Benjamin Murauer, Günther Specht
Universität Innsbruck, Austria
manfred.moosleitner@uibk.ac.at,b.murauer@posteo.de,guenther.specht@uibk.ac.at
ABSTRACT We included two additional features that were not in the JSON
This paper summarizes the contribution of our team UIBK-DBIS- files directly. Firstly, we crawled all URLs which were included in
FAKENEWS to the task “FakeNews: Corona virus and 5G conspir- the messages and extracted the content of the sites tag,
acy” as part of MediaEval 2020. The goal for this task is to classify hoping that it would contain a distinctive vocabulary. Secondly, we
tweets as “5G corona virus conspiracy”, “other conspiracy”, or “non used the free OCR software tesseract 2 to find any text within the
conspiracy”, based on text analysis and based on the retweet graphs. images that are included in the messages.
We achieved our best results using a calibrated linear SVM with We tested linear support vector machines and extra random
word and character n-grams for the text classification task and trees as classifiers, and also added the option of calibrating the SVM
a non-calibrated linear SVM with graph statistics for the graph using Platt’s method [7]. These classifiers have been well-studied
classification task. and perform well in diverse text classification tasks [10], and can
compete with neural-network-based approaches in many fields like
spam detection [5].
1 INTRODUCTION
The main objective in the task is to distinguish tweets and classify 2.2 Subtask 2: Retweet-Follower-Graphs
them as either (1) contributing to a conspiracy suggesting that the
Standard graph statistics like the number of nodes or the graphs
5G network technology caused the SARS-CoV-2 virus epidemic,
degrees are known to carry characteristics about the retweet graph
(2) contributing to a different conspiracy, or (3) not contribute to
to help in classification [1]. Also, algorithms like HITS [3] and
a conspiracy. For the first subtask, this classification is based on
PageRank [6] could produce discriminating features, as they were
the text content of the tweets. The second subtask focuses on the
used on retweet graphs by Yang et al. in [11] to distinguish between
retweet and follower graph of the tweets. A detailed description
tweets that are interesting only to a small group of people or a
and the results of the challenge can be found in [8], the collection
broader audience. Thus, we used the statistical networking Python
of the data is described in [9].
package NetworkX 3 to extract statistical figures describing the
In the remainder of this overview, we present our solutions for
retweet-follower-graphs. For the first run of the second subtask, we
the two subtasks in the following Section 2, and discuss the results
calculate order, size, degree, indegree, outdegree, number of connected
thereafter in Section 3.
components, density, transitivity, pagerank, HITS (hubs, authorites),
number of partitions, planarity, and number of cycles, and combined
2 METHODOLOGY them into a single feature vector.
In both subtasks, the participants are allowed to submit 5 different Some of the functions in NetworkX to calculate the graph sta-
solutions, whereas the first 2 solutions of each subtask are restricted tistics return lists of variable length, as their number depends on
to only use part of the information available. In the remaining 3 the number of nodes and edges. To create fixed-length feature vec-
submissions, also external data points may be used. tors, we computed arithmetic mean, standard deviation, and the
five-number summary of the values in the individual lists, and used
2.1 Subtask 1: Twitter Messages these as features. For the second run in subtask 2, we additionally
We extract character and word-based 𝑛-grams from the text of used the data from the nodes files, from which we calculated min,
the tweets and use them as features for our classification models. max, mean, and standard deviation of the number of friends and
This has been shown to be effective and versatile in different text followers, and added these to the feature vectors calculated for the
classification task ranging from stance detection [2] to classifying first run.
hacked tweet accounts [4]. We tested different parameters in a grid 2 https://tesseract-ocr.github.io/
search, the values of which are listed in Table 1. 3 https://networkx.org/
Submissions 2 may include additional information, so we added
all features that were included in the JSON structure, which corre-
Table 1: Hyperparameters tested in grid search.
spond to the fields available from Twitter’s API1 . We transformed
all textual features to tf/idf normalized frequencies of 𝑛-grams,
as listed in Table 1, left the numeric features were left as-is, and Parameter Tested values
mapped all categorical features to one-hot vectors. Word & character 𝑛-gram size1 [1,2,3,4]
1 https://developer.twitter.com/en/docs/twitter-api SVM: C [0.1, 1, 10]
Extra Trees: number of trees [1, 2, 3, 4] ×103
Copyright 2020 for this paper by its authors. Use permitted under Creative Commons Poly. degree [2, 3]
License Attribution 4.0 International (CC BY 4.0). Poly. include bias [True, False]
MediaEval’20, 14-15 December 2020, Online KNN: number of neighbors [3, 4, 5, 10, 20, 50]
MediaEval’20, 14-15 December 2020, Online M. Moosleitner, B. Murauer, G. Specht
Table 2: Evaluation results measured with Matthew’s corre-
Figure 1: Top 3 positive and negative SVM coefficients for lation coefficient.
each class after fitting the message bodies of the training
data. Phase Model Run 1 Run 2
5G corona conspiracy Linear SVM (calibrated) 0.432 0.412
conspiracies 5g Training Linear SVM 0.428 0.404
better wuhan Extra Random Trees 0.274 0.253
burning symptoms
Evaluation Linear SVM (calibrated) 0.440 0.441
−8 −6 −4 −2 0 2 4 6 8
(a) Results of Subtask 1
No conspiracy
5g burning Phase Model Run 1 Run 2
wuhan facebook
Linear SVM (calibrated) 0.003 0.054
symptoms conspiracies
Linear SVM 0.127 0.197
−8 −6 −4 −2 0 2 4 6 8 Training KNN 0.118 0.135
Other conspiracy Extra Random Trees 0.089 0.091
body cancer Gaussian Naive Bayes 0.092 0.101
because but Evaluation Linear SVM 0.090 0.092
already msm
(b) Results of Subtask 2
−8 −6 −4 −2 0 2 4 6 8
Table 3: Best parameters for the four submissions.
Since we extracted significantly fewer features in the second
subtask, we added polynomial feature generation, and added a
gaussian naïve Bayes classifier and a K-nearest neighbor to the Subm. Parameters
models from the first subtask. Both are well-studied algorithms and Text 1 word-1-grams + character-3+4-grams, calibrated SVM, C=0.1
we were interested in how well they would perform for this task. Text 2 word-1-grams + character-3+4-grams, calibrated SVM, C=0.1
We tested several parameters in a grid search, which are displayed Graph 1 linear SVM, C=10, Poly. deg=2, Poly. include bias = True
in Table 1. Graph 2 linear SVM, C=10, Poly. deg=3, Poly. include bias = False
3 RESULTS AND DISCUSSION
After preliminary experiments for both subtasks, we selected the discussing the telecommunication standard. This relationship could
setup with the highest MCC score in a 10-fold cross-validation be experimented with in more detail using topic modeling.
setup as the model that predicts our submission results for each
subtask. 3.2 Subtask 2
Similar to subtask 1, we used grid search to find the best performing
3.1 Subtask 1 classifier and parameters. The scores of the classifiers were rather
The scores displayed in Table 2a show that the SVM model clearly similar, with the linear SVM producing the best score with the
outperforms the extra random trees approach in the first subtask. parameters C=10. While using polynomial features at all increased
Thereby, calibrating the SVM increased the performance slightly. the result in both submissions by 0.05, whereas the parameters
Interestingly, the performance of the classifiers dropped when (degree=[2,3], include bias=[true, false]) did not have a great influ-
taking more features into account for the second submission. This ence (< 0.01 MCC). as shown in Table 3. The results in training and
indicates that either too many features are extracted from the text, evaluation approaches for subtask 2 were quite low, as displayed
or that the additional meta-information was not expressive to the in Table 2b. Interestingly, our MCC validation scores for subtask
problem. Nevertheless, we submitted the two results in this state, 2 were lower than the training scores, which is in contrast to the
being aware that we could have possibly increased the performance scores of subtask 1, where the validation scores were slightly better
of the second submission by ignoring the meta-features. The evalua- than our training scores.
tion results, on the other hand, don’t display a performance decrease
between the two submissions, where both runs result in a score of 4 CONCLUSION
0.440 and 0.441, respectively. As shown in Table 3, the best results Our simple text-based approaches were able to classify the tweets
were obtained by combining word unigrams and character-3- and reliably, and the coefficients of the model give insights into the
-4-grams and a strict regulation parameter of C=0.1. most important terms. We suggest that more preprocessing might
Using a linear SVM as a model allows an easy interpretation of further improve these results.
the importance of words by looking at the respective coefficients. The simple graph statistics, on the other hand, were not expres-
For each output class, Figure 1 shows the terms with the three sive enough for this task. Here, incorporating more metadata like
highest and lowest coefficients. The high value for the term 5g the time between the retweets might improve the classification
suggests that not many topics within the other conspiracies are results.
FakeNews: Corona virus and 5G conspiracy MediaEval’20, 14-15 December 2020, Online
REFERENCES
[1] David R Bild, Yue Liu, Robert P Dick, Z Morley Mao, and Dan S Wallach.
Aggregate characterization of user behavior in twitter and analysis of
the retweet graph. ACM Transactions on Internet Technology (TOIT),
15(1):1–24, 2015.
[2] Peter Bourgonje, Julian Moreno Schneider, and Georg Rehm. From
clickbait to fake news detection: an approach based on detecting the
stance of headlines to articles. In Proceedings of the 2017 EMNLP
Workshop: Natural Language Processing meets Journalism, pages 84–89,
2017.
[3] Jon M Kleinberg. Hubs, authorities, and communities. ACM computing
surveys (CSUR), 31(4es):5–es, 1999.
[4] Benjamin Murauer, Eva Zangerle, and Günther Specht. A peer-based
approach on analyzing hacked twitter accounts. In Proceedings of the
50th Hawaii International Conference on System Sciences, 2017.
[5] N. L. Octaviani, E. Hari Rachmawanto, C. A. Sari, and D. Rosal Ignatius
Moses Setiadi. Comparison of multinomial naïve bayes classifier,
support vector machine, and recurrent neural network to classify email
spams. In 2020 International Seminar on Application for Technology of
Information and Communication (iSemantic), pages 17–21, 2020.
[6] Lawrence Page, Sergey Brin, Rajeev Motwani, and Terry Winograd.
The pagerank citation ranking: Bringing order to the web. Technical
report, Stanford InfoLab, 1999.
[7] John Platt. Probabilistic outputs for support vector machines and com-
parisons to regularized likelihood methods. Advanced Large Margin
Classifiers, 10, June 2000.
[8] Konstantin Pogorelov, Daniel Thilo Schroeder, Luk Burchard, Johannes
Moe, Stefan Brenner, Petra Filkukova, and Johannes Langguth. Fak-
enews: Corona virus and 5g conspiracy task at mediaeval 2020. In
MediaEval 2020 Workshop, 2020.
[9] Daniel Thilo Schroeder, Konstantin Pogorelov, and Johannes Langguth.
Fact: a framework for analysis and capture of twitter graphs. In 2019
Sixth International Conference on Social Networks Analysis, Manage-
ment and Security (SNAMS), pages 134–141. IEEE, 2019.
[10] Simon Tong and Daphne Koller. Support vector machine active learn-
ing with applications to text classification. Journal of machine learning
research, 2(Nov):45–66, 2001.
[11] Min-Chul Yang, Jung-Tae Lee, Seung-Wook Lee, and Hae-Chang Rim.
Finding interesting posts in twitter based on retweet graph analysis. In
Proceedings of the 35th international ACM SIGIR conference on Research
and development in information retrieval, pages 1073–1074, 2012.