=Paper=
{{Paper
|id=Vol-2696/paper_225
|storemode=property
|title=Profiling Spreaders of Disinformation on Twitter: IKMLab and Softbank Submission
|pdfUrl=https://ceur-ws.org/Vol-2696/paper_225.pdf
|volume=Vol-2696
|authors=Timothy Niven,Hung-Yu Kao,Hsin-Yang Wang
|dblpUrl=https://dblp.org/rec/conf/clef/NivenKW20
}}
==Profiling Spreaders of Disinformation on Twitter: IKMLab and Softbank Submission==
Profiling Spreaders of Disinformation on Twitter: IKMLab and SoftBank Submission Notebook for PAN at CLEF 2020 Timothy Niven, Hung-Yu Kao, and Hsin-Yang Wang Intelligent Knowledge Management Lab National Cheng Kung University, Tainan, Taiwan and AI Strategy Office, SoftBank Corp., Tokyo, Japan tim.niven.public@gmail.com, hykao@mail.ncku.edu.tw, hsinyang.wang@g.softbank.co.jp Abstract The problem we address is classifying whether a Twitter user has spread confirmed disinformation or not. We used two types of features that had validity in the training set: features that indicate thoughtfulness, and features re- flecting emotional states. We attempted to capture thoughtfulness via the rate of function word usage and constituency tree features reflecting sentence complex- ity. We added features for sentiment in general and negative sentiment in partic- ular to measure emotional arousal. We also experimented with custom lexicons for anger and distrust. Our classifier was an ensembled Support Vector Classifier, Random Forest, and Naive Bayes algorithms. We only considered the English data. Our cross-validated training set accuracy was 89.3%, but significantly over- fit the training data, achieving 61.0% test set accuracy. 1 Introduction The “Profiling spreaders of fake news” task at PAN 2020 [15] supplies 100 tweets per 300 users, where half of the users are confirmed to have spread deliberate disinforma- tion.1 The tweets are almost entirely retweets and shared news headlines. Task partic- ipants are required to develop techniques to separate the spreaders of disinformation from the controls. The task authors provide English and Spanish datasets to facilitate multilingual techniques. This is an exciting contribution, as solutions that work across languages are likely to reflect learning about the task, as opposed to “solving datasets” (i.e. overfitting the bias distribution shared across training and held-out testing sets, e.g. as a result of the data collection process) [9]. Copyright c 2020 for this paper by its authors. Use permitted under Creative Commons Li- cense Attribution 4.0 International (CC BY 4.0). CLEF 2020, 22-25 September 2020, Thessa- loniki, Greece. 1 Following Taiwan’s Digital Minister, Audrey Tang, we prefer the term “disinformation,” as “fake news” carries connotations that can undermine journalism [4]. 2 Our Approach Regrettably, for our submission we only had a chance to work on the English data. Our approach was to restrict ourselves to considering how the tweets discussed their topics, and not what they were talking about. Although the dataset comprises 30,000 tweets there are only 300 labels.2 We were concerned that a solution such as BERT [6] might focus on keywords that have statistical validity across both train and test, but nevertheless overfit the dataset as a whole [11]. We were also interested to find features that help us understand the phenomenon. We therefore consider predictive accuracy as important but not all-encompassing, and favour solutions that are interpretable.3 The features we chose are designed to measure thoughtfulness and emotional arousal. Measurements were made from lexicons and constituency tree parses. For lexical cate- gories, we counted the frequency of words for each tweet, and then took the mean over tweets per user. Formally, let the set of words in category c be Lc . For each user, u, (u) tokenize all 100 tweets yielding token vectors ti , i ∈ [0, 100], and count the propor- tion of words belonging to category c. Then average over all user tweets to get the final frequency of usage per user: P 1 X t∈ti(u) 1[t ∈ Lc ] 100 (u) fc = (u) . 100 i=1 |t | i Constituency tree features were obtained with the Berkeley Neural Parser [10] and will be detailed below. 2.1 Thoughtfulness We reasoned that people who share more thoughtful content are less likely to spread disinformation. We developed two sets of features to measure and investigate this hy- pothesis: (1) the frequency of function word usage; (2) constituency tree features. Function Word Usage Rates Higher usage rates of function words indicate more com- plex sentences, and therefore may be seen to express thinking effort. The training data shows some support for the existence of a threshold of function word usage rate above which a user is unlikely to have spread disinformation. We use the open-source English function word lists provided by [12], and use the categories: personal pronouns, adverbs, and all function words. Distributions of fc for these categories are given in Figure 1. We included all three features in our final clas- sifier as cross-validation indicated they were useful, however the adverbs and personal pronouns are questionable features. The average number of tokens in a tweet after our tokenization method is just over 13. Yet the baseline usage rate for adverbs is 0.06 for adverbs and 0.04 for personal pronouns. Therefore, the tweets are not long enough to yield reliable statistics for these word categories. The complete function words category 2 The labels apply to the 300 authors, each of which contribute 100 tweets to the dataset. 3 In this respect, our choice to ensemble with XGBoost was questionable. 7 Not Spreader Spreader 35 20 6 30 5 25 15 4 20 10 3 15 10 2 5 5 1 0 0 0 0.00 0.05 0.10 0.15 0.00 0.05 0.10 0.15 0.0 0.2 0.4 0.6 Adverbs Personal Pronouns Function Words Figure 1. Distributions of usage rates of function word categories. is more reliable, with baseline frequency of 0.33. There appears to be a threshold for function word usage rates above which the probability of spreading disinformation is low in this dataset. This threshold appears to be quite high compared to the rest of the distribution, suggesting this is not a feature with great power to solve this task on its own. Constituency Tree Features Following the same intuition that sentence complexity indicates thoughtfulness, we investigated related features from constituency tree parses. Specifically, we calculated the average branching factor, and highest noun and verb phrases in each constituency tree. We consider a tweet as a single sentence, which is reasonable given that most tweets are headlines of news articles. From the constituency trees we extract the average branching factor, and average maximum noun and verb (u) phrases as follows. Let Ni be the set of internal non-leaf nodes in the constituency tree for tweet i from user u. Let φbf (·) count the number of children of a node. Then, 100 (u) 1 X 1 X µbf = φbf (n) . 100 i=1 |N (u) | (u) i n∈Ni (u) For noun and verb phrases, let Pi be the set of sub-trees representing either type of (u) phrase, and let φhP (·) select the height of the tallest sub-tree in Pi of phrase type P . Then, 100 (u) 1 X (u) µhNP = φh (P ) 100 i=1 NP i 100 (u) 1 X (u) µhVP = φh (P ). 100 i=1 VP i 2.00 0.7 Not Spreader Spreader 1.75 0.4 0.6 1.50 0.5 0.3 1.25 0.4 1.00 0.3 0.2 0.75 0.2 0.50 0.1 0.25 0.1 0.00 0.0 0.0 2 4 6 8 10 2 4 6 8 0 2 4 6 8 10 Average Branching Factor Max NP Height Max VP Height Figure 2. Distributions of constituency tree features. The distributions over both labels are given in Figure 2. Once again the distributions are not very well separated, but indicate some weak support for our hypothesis that higher average branching factor and higher average noun phrases are more associated with the control group above a (fairly high) threshold. Interestingly, the pattern is opposite to what we expected for verb phrases, requiring further investigation. 2.2 Emotional Arousal We initially investigated emotional arousal using the Linguistic Inquiry and Word Count (LIWC) dictionary [14]. However, due to being proprietary, we could not package it into our final software solution. We are also advocates of open science and would like to investigate alternatives to drive scientific progress, and provide resources equally available to less privileged researchers. Nevertheless, the value of the LIWC dictionary is beyond doubt, and still appears stronger than the alternatives we tried.4 Our LIWC analysis indicated the usefulness of: anger, negative emotion, and anx- iety. The finding that anger is useful is consistent with previous work [8]. Using these three features, a support vector classifier could achieve 70% cross-validated accuracy on the training set. We therefore set ourselves the task of finding an equally good open source lexicon. We did not find one for anxiety, and so skipped that category. SentiWordNet For negative emotions, we settled on SentiWordNet [1]. Figure 3 shows the distributions for sentiment (of both kinds), and negative sentiment. An unexpected result is that more sentiment is associated with the control group, however negative sentiment shows the expected pattern. Again, these features appear to have relatively weak discriminative power, especially compared to the LIWC lexicon. 4 One exception is the function word categories, for which the lexicon from [12] appears to yield very similar distributions. This is consistent with their findings, obtained on a different task and dataset. Our findings in this respect therefore give extra weight to the claim that this is a good open source alternative for the LIWC function word lexicon. 80 Not Spreader Spreader 60 70 50 60 40 50 40 30 30 20 20 10 10 0 0 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 0.00 0.02 0.04 0.06 0.08 0.10 Sentiment Negative Sentiment Figure 3. Distributions of sentiment and negative sentiment features. Anger Of all the LIWC features, anger provided the clearest separation between the two classes, consistent with our understanding of how much disinformation works. Anger is known as an emotion that leads to action (REF). Therefore, if disinforma- tion can bring someone to feel anger and outrage, it can promote its spread.5 We did not find an open-source offering that could separate the two classes like LIWC does.6 Given the importance of this feature, our broader goal of finding alternatives to LIWC, and general interest, we decided to experiment with making one. Our anger dictionary is far from a thorough or complete work. Our only criteria for success was a distribution that approximated the LIWC anger lexicon on this data. Since we used the training data to tune the lexicon, it also very likely overfits. Inspired by previous work, we decided to exploit emoji on Twitter [2]. We used the twint library7 to obtain tweets containing exclusively anger or joy emoticons in multiple 6-month time bins from the middle of 2020 to the end of 2014. We collection 20,000 tweets per emotion per time bin. We collected joy as a control, since in the average it is unlikely anger words will be present in joyous posts. The point of collecting over different times was to try and factor out the content words associated with anger at different times. We found “Trump” to be the most popular correlate of anger going back to before 2016, requiring a large number of timesteps to push him down our ranked list of anger words. Utilizing a simple technique from social scientific research into media bias [7], we ranked words from this dataset by their χ2 statistic comparing the frequency of each word’s appearance in both anger and joy tweets. To account for variation over time, we scaled the χ2 statistic by its entropy over time periods, promoting stable anger words, and demoting words whose correlation with each emotion is variable over time. The final results appeared quite reasonable, with the top five words being: angry, bloody, crap, disgusted, and unfair. We tuned the lexicon to the dataset by taking the top-50 5 Taiwan’s strategy for combating disinformation, “Humor over Rumor,” is based on this under- standing: humor being the best antidote to anger [16]. 6 Our search is unlikely to have been exhaustive. 7 https://github.com/twintproject/twint. 70 250 Not Spreader Spreader 60 200 50 150 40 30 100 20 50 10 0 0 0.01 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.005 0.000 0.005 0.010 0.015 0.020 0.025 0.030 Anger Distrust Figure 4. Distributions of anger and distrust features. stems that were predictive of the “Spreader” class, as determined by average weight given to each word of a linear support vector classifer using 5-fold cross-validation. The resulting distribution of applying this lexicon is given in Figure 4. It does a much better job of distinguishing the two classes. Distrust One recurrent theme in contemporary disinformation has been distrust in in- stitutions. We therefore experimented with a creating a “distrust” lexicon. We used the χ2 statistics of unigrams comparing their frequencies in the two class labels, and man- ually inspected the word list, pulling out words pertaining to distrust. The final lexicon includes 36 words such as: fake, suspect, unbelievable, doubt, question, lie, and scam. The resulting distributions are show in Figure 4. We generally see higher distrust scores for spreaders of disinformation, suggesting this could be a promising feature. 3 Results We use NLTK [3] to tokenize the tweets, removing the special characters introduced by the dataset authors, such as “#URL#.” The features described above are then calculated for each tweet, and averaged over users, as given in the equations above. We separately trained support vector, random forest, and naive Bayes classifiers, us- ing 5-fold cross-validation to find optimal hyperparameters, using scikit-learn [13]. The support vector classifier used an RBF kernel and C = 1., achieving 69.7% accuracy on the training set. The random forest uses 40 estimators with a max depth of 5, achieving 68.3% accuracy. The naive Bayes achieved 64.3% accuracy. We then ensembled these classifiers with XGBoost [5], which achieved 5-fold cross-validated accuracy of 89.3% on the training set. 4 Discussion It is very possible that the minor distributional differences in many of our features are an artifact of random sampling and fail to generalize, particularly those relating to thought- fulness. However, cross-validation indicated their usefulness for this task, and the intu- itions behind many of the features seem reasonable. It is likely our use of XGBoost has lead to overfitting due to the small number of labeled data points. We may also have overfit the training set when building our anger and distrust lexicons. It would be much more interesting to compare these features across languages. Per- haps due in part to our unfamiliarity with Spanish, we were unable to easily find open- source resources to match what we have obtained for English. References 1. Baccianella, S., Esuli, A., Sebastiani, F.: SentiWordNet 3.0: An enhanced lexical resource for sentiment analysis and opinion mining. In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC’10). European Language Resources Association (ELRA), Valletta, Malta (May 2010), http: //www.lrec-conf.org/proceedings/lrec2010/pdf/769_Paper.pdf 2. Bandhakavi, A., Wiratunga, N., P, D., Massie, S.: Generating a word-emotion lexicon from #emotional tweets. In: Proceedings of the Third Joint Conference on Lexical and Computational Semantics (*SEM 2014). pp. 12–21. Association for Computational Linguistics and Dublin City University, Dublin, Ireland (Aug 2014). https://doi.org/10.3115/v1/S14-1002, https://www.aclweb.org/anthology/S14-1002 3. Bird, Steven, L.E., Klen, E.: Natural Language Processing with Python. O’Reilly Media Inc. (2009) 4. Butler, S., Hsu, I.: Q&A: Taiwan’s digital minister on combatting disinformation without censorship. Committee to Protect Journalists (2019), https://cpj.org/2019/05/ qa-taiwans-digital-minister-on-combatting-disinfor/ 5. Chen, T., Guestrin, C.: Xgboost. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (Aug 2016). https://doi.org/10.1145/2939672.2939785, http://dx.doi.org/10.1145/2939672.2939785 6. Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. CoRR abs/1810.04805 (2018), http://arxiv.org/abs/1810.04805 7. Gentzkow, M., Shapiro, J.: What drives media slant? evidence from u.s. daily newspapers. Econometrica 78, 35–71 (12 2006). https://doi.org/10.2139/ssrn.947640 8. Ghanem, B., Rosso, P., Rangel, F.: An Emotional Analysis of False Information in Social Media and News Articles. ACM Transactions on Internet Technology (TOIT) 20(2), 1–18 (2020) 9. He, H., Zha, S., Wang, H.: Unlearn dataset bias in natural language inference by fitting the residual (2019) 10. Kitaev, N., Klein, D.: Constituency parsing with a self-attentive encoder. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics, Melbourne, Australia (July 2018) 11. Niven, T., Kao, H.Y.: Probing neural network comprehension of natural language arguments. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. pp. 4658–4664. Association for Computational Linguistics, Florence, Italy (Jul 2019). https://doi.org/10.18653/v1/P19-1459, https://www.aclweb.org/anthology/P19-1459 12. Noble, B., Fernández, R.: Centre stage: How social network position shapes linguistic coordination. In: Proceedings of the 6th Workshop on Cognitive Modeling and Computational Linguistics. pp. 29–38. Association for Computational Linguistics, Denver, Colorado (Jun 2015). https://doi.org/10.3115/v1/W15-1104, https://www.aclweb.org/anthology/W15-1104 13. Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., Duchesnay, E.: Scikit-learn: Machine learning in Python. Journal of Machine Learning Research 12, 2825–2830 (2011) 14. Pennebaker, J., Boyd, R., Jordan, K., Blackburn, K.: The development and psychometric properties of liwc2015 (09 2015). https://doi.org/10.15781/T29G6Z 15. Rangel, F., Giachanou, A., Ghanem, B., Rosso, P.: Overview of the 8th Author Profiling Task at PAN 2020: Profiling Fake News Spreaders on Twitter. In: Cappellato, L., Eickhoff, C., Ferro, N., Névéol, A. (eds.) CLEF 2020 Labs and Workshops, Notebook Papers. CEUR Workshop Proceedings (Sep 2020), CEUR-WS.org 16. Silva, S.: Coronavirus: How map hacks and buttocks helped taiwan fight covid-19. BBC (2020), https://www.bbc.com/news/technology-52883838