=Paper=
{{Paper
|id=Vol-3180/paper-46
|storemode=property
|title=FoSIL at CheckThat! 2022: Using Human Behaviour-Based Optimization for Text Classification
|pdfUrl=https://ceur-ws.org/Vol-3180/paper-46.pdf
|volume=Vol-3180
|authors=Andy Ludwig,Jenny Felser,Jian Xi,Dirk Labudde,Michael Spranger
|dblpUrl=https://dblp.org/rec/conf/clef/LudwigFXLS22
}}
==FoSIL at CheckThat! 2022: Using Human Behaviour-Based Optimization for Text Classification==
FoSIL at CheckThat! 2022: Using Human Behaviour-Based Optimization for Text Classification Andy Ludwig1 , Jenny Felser1 , Jian Xi1 , Dirk Labudde1 and Michael Spranger1 1 University of Applied Sciences Mittweida, Technikumplatz 17, 09664 Mittweida, Germany Abstract Nowadays, a huge amount of information and news articles are available every day. The events of recent years have shown that Fake News can severely shake trust in politics and science. Unfortunately, a decision can only be made about the truthfulness of a fraction of all news and posts. In this respect, the CLEF2022-CheckThat! shared task 3a adresses this problem. In this paper, we propose a new classification approach using a novel metaheuristic feature selection algorithm that mimics human behavior. The results show that the performance of a baseline classifier can achieve higher performance by combining with this algorithm with only a fraction of the features. Keywords fake news detection, text classification, feature selection, human behavior-based optimization 1. Introduction In times of constant availability of vast amounts of information, people have to judge the truth of news in a short time. This assessment is often neglected due to the fast-moving nature of the news. There are various reasons why authors, whether intentionally or unintentionally, contribute to the generation of untrustworthy content. In particular, sources that deliberately disseminate false information pose a danger to consumers of news. Effective methods for detecting fake news are essential in the fight against the targeted spread of fake news. Continuous research and development of approaches to detect misinformation are highly important. This is one mission of the CLEF2022 - CheckThat! Lab [1, 2]. In general, the Lab’s goal is to verify the veracity of claims. Task 3a takes up the challenge of assessing the truth content of news articles [3]. This paper presents a novel approach for text classification. The concept is based on human behaviour-based optimization (HBBO) [4]. This meta heuristic optimization approach uses some fundamental interactions and behaviours of humans. The potential of this adaptation was used for the fake news detection in task 3a. The paper is structured as follows: section 2 presents related works, section 3 describes the human behaviour-based optimization, section 4 summarizes the adaption of the optimization approach for text classification, section 5 presents the given data and the conducted experiments, in section 6 the results are discussed and finally in section 7 a conclusion is given. CLEF 2022: Conference and Labs of the Evaluation Forum, September 5–8, 2022, Bologna, Italy $ ludwig1@hs-mittweida.de (A. Ludwig); jfelser@hs-mittweida.de (J. Felser); xi@hs-mittweida.de (J. Xi); labudde@hs-mittweida.de (D. Labudde); spranger@hs-mittweida.de (M. Spranger) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) 2. Related Work The assessment of the truthfulness of news or claims is a special case of text classification whose importance has increased greatly in recent times. Especially during the Covid-19 pandemic situation a larger group of people was challenged to judge news correctly. This situation makes the invention of automatic text classification systems necessary [5]. To pool a lot of competence and to push the development forward, a task for different participants can be done like the annual task CheckThat! Lab since 2018. The results of last year are summarized in [6]. These tasks need a high quality dataset [7], which is also reflected in the available labels [8]. To improve the results for fake news detection and the related text classification task, a wide range of possible approaches can be explored. In this paper, the research focus was on a new technique to select the best features for training. To solve these optimization tasks, nature inspired meta-heuristic techniques can be applied. An overview of already known approaches is given in [9]. For example, the ant colony optimization is a possible algorithm for feature selection in text classification tasks [10]. The nature-inspired algorithm used in this paper is derived from human behaviour. The basis of this approach was presented in [4]. This algorithm was used already in [11] in combination with self-organizing maps in context of cryptanalytic resistance. The authors compare this approach with other meta-heuristic solutions like ant colony optimization. Next to the approach of [4], other algorithms based on human behaviour were described. Firstly, in [12] the goal is to solve optimization tasks by simulating the phases of knowledge gaining and sharing in younger and older years of human life. A different aspect of optimization algorithms under usage of human behaviour is presented in [13]. This approach focuses on the adaption of behaviours and manners of other humans for example in family structures. 3. Human Behaviour-Based Optimization (HBBO) In this section, we briefly describe Ahmadi’s novel swarm intelligence-based optimization approach [4] considering human behavior, which forms the basis for the feature selection approach used here. A central aspect in all phases of the algorithm is an individual’s pursuit of self-optimization at different stages of his or her life. At the same time, different individuals have different levels of experience in their field, and some of them become experts in one of them (e.g., art, music, or science, etc.). The optimization of individual performance is carried out iteratively in four main steps: Ini- tialization, Education, Consultation, and Field Changing. In the Initialization step the population is built. Each individual in the population is assigned an area of interest in which improve- ment should be achieved. Depending on the underlying optimization problem, an individual )︀𝑇 is represented as a vector of characteristic variables (features). 𝐼 = 𝑥1 𝑥2 . . . 𝑥𝑁 . An (︀ expert is determined for each field. The expert provides the best function value depending on his individual feature set. The formal definition of an expert is shown in equation (1), where 𝐼 denotes a person, 𝐸 denotes an expert, and 𝐹 denotes the actual field of expertise. 𝐼 (𝐹 ) → 𝐸 (𝐹 ) = argmin 𝑓 error (𝐹 ) (1) The first step in the improvement process, is Education. Individuals in each area learn from the respective expert. The learning process comprises the improvement of the individual characteristic values by those of the expert and aims at the reduction of their own error function value. A similar procedure is used in Consultation. In this stage, individuals can learn from any other individual, not only from an expert. For this purpose, some variables are merged between two randomly selected individuals. The consultation is called effective and the merged set of variables is kept if the updated variables lead to better function values. Otherwise, the update is reversed. The last step is Field Changing. Whether an individual changes his field of interest is calculated using a rank probability method and the roulette wheel selection. After Initialization, the three remaining steps are repeated until a stop criterion is met, e.g. if the average function values do not change (or change too little) or the number of iterations reaches a maximum. This algorithm can be adapted to find an optimal feature set for text classification. The next section explains the adaptation in detail. 4. Adaptation for Text Classification In order to adapt the HBBO approach for feature selection, the optimization objective, i.e., an objective function, must be determined. Here, the 𝐹1 -score is chosen on a test set. For this purpose, the preprocessed dataset must be divided into training and test data. To obtain reliable results and avoid overfitting, the data set is split in terms of k-fold cross-validation. Each fold results in an optimal feature set in the form of a document term matrix (DTM) that is able to classify the respective test data with the greatest possible success. Finally, the results of all the folds are merged into a DTM that contains only the most successful features to classify the entire dataset. For assessing the performance after each optimization step, a support vector machine (SVM) is used. 4.1. Initialization During Initialization, each input document is considered as an single individual and encoded in the form of a bit vector. This vector represents an individual’s knowledge. Each individual is assigned the class of the respective document, which also represents the field of interest. All vectorized documents together form the entire population. In contrast to the original algorithm where each individual is optimized, here subsets of individuals are optimized together. This leads to a smaller set of documents achieving good or even better classification results than the original set of all documents. A group of individuals contains all fields, which leads to a simultaneous optimization of all classes within a group. For this purpose, the individuals are grouped into subsets, with each subset equally populated with individuals of each class (stratified approach). Each group of individuals is optimized separately. The size of the subsets is a hyperparameter. The remaining three steps are applied iteratively to each group of each fold. The number of iterations needs to be specified as a hyperparameter. 4.2. Education In the Education step, a group of individuals must first be determined for each field that achieves the lowest error with respect to the objective function on the test set according to equation (2). Each of these groups is considered to be the expert group for that particular field. 𝐼 (𝐹 ) → 𝐸 (𝐹 ) = argmax 𝐹1 (2) Subsequently, a subset 𝑆𝐸 of features of the expert group, considering all classes, is merged with the features of other individuals, belonging to the remaining groups. This procedure corresponds to the non-experts approaching the expert group by updating their feature vector. The number of terms transferred in this step is another hyperparameter, as well as the number of adapted individuals, i.e., documents. Equation (3) shows the formal definition of this step, (𝐹 ) where 𝐼𝑖 denotes the 𝑖𝑡ℎ individual 𝐼 in the specific field 𝐹 and 𝑆𝐸 ⊆ 𝐸 (𝐹 ) . 𝐼𝑖 ∪ 𝑆𝐸 , if𝑓 error (𝐼𝑖 ∪ 𝑆𝐸 ) < 𝑓 error (𝐼𝑖 ) {︃ (𝐹 ) (𝐹 ) (𝐹 ) (𝐹 ) 𝐼𝑖 = (𝐹 ) (3) 𝐼𝑖 , otherwise The feature update is considered successful and the merged feature set is retained only if the individual achieves an improvement with respect to the objective function, i.e., better classification results. Otherwise, the previous feature set remains unchanged. 4.3. Consultation In Consultation, the basic procedure is very similar to that in the Education step. The main difference is in the group of individuals from which the features for merging are taken. Regardless of expert status, features are merged between two randomly selected groups of individuals. This leads to greater heterogeneity of terms across all groups. Equation (4) shows the formal (𝐹 ) definition of Consultation, where 𝑆𝑗 is a subset of all features of individual 𝐼𝑗 and 𝑆𝑗 ⊆ 𝐼𝑗 . 𝐼𝑖 ∪ 𝑆𝑗 , if𝑓 error (𝐼𝑖 ∪ 𝑆𝑗 ) < 𝑓 error (𝐼𝑖 ) {︃ (𝐹 ) (𝐹 ) (𝐹 ) (𝐹 ) 𝐼𝑖 = (𝐹 ) (4) 𝐼𝑖 , otherwise As in Education, the updated feature set is retained only if the 𝐹1 -score improves; otherwise, the previous terms remain unchanged. Again, the number of terms exchanged and the number of individuals paired can be controlled using hyperparameters. 4.4. Field changing The last step - Field Changing - does not manipulate the individuals’ terms, but changes the field associated with them. The number of randomly selected individuals changing the field is another hyperparameter. In a multilevel classification, an individual can simply change their area of interest to that of the most successful expert group. For this purpose, all expert groups are ordered by their 𝐹1 -score, as shown in equation (5), where 𝐸 (𝐹𝑥 ) is the expert group in the respective area. {︁ (︁ )︁}︁ 𝑅 = 𝐸 (𝐹1 ) , ..., 𝐸 (𝐹𝑛 ) | 𝑓 error 𝐸 (𝐹1 ) < 𝐸 (𝐹2 ) < ... < 𝐸 (𝐹𝑛 ) (5) Afterwards, the expert group with the highest rank in 𝑅 determines the field of the individual willing to switch, as shown in equation (6). {︁ (︁ )︁}︁ 𝐹𝑥* = 𝐹𝑥 |𝐸 (𝐹𝑥 ) = argmin𝐸 (𝐹𝑥 ) ∈𝑅 𝑓 𝑒𝑟𝑟𝑜𝑟 𝐸 (𝐹𝑥 ) (6) In the special case of a binary classification, the process simply reduces to a field change, i.e., an inversion of the class label. Again, the new field is retained only if it leads to an improvement in the objective function value. 5. Experiments and Results In this work, the adaptation of HBBO, as discussed in section 4, was applied to the problem of detecting fake news in the context of the CLEF2022-CheckThat! shared task 3a. The documents provided were news articles in English, which were to be grouped into the classes true, false, partially false and other with regard to potentially containing fake news. For training the model, the provided training and development sets were combined so that the total training corpus comprised 1, 264 documents, each of which was assigned to exactly one of the four mentioned classes. The resulting class distribution of the training corpus is shown in Figure 1. False True 211 578 117 Other 358 Partially False Figure 1: Distribution of classes and documents in the corpus of training data. Before the HBBO algorithm can be applied, the input data must be cleaned. To this end, a wide range of cleaning steps were performed: • Combination of article and headline into one pseudo document, • Replacement of newlines and tabs with white space, • Removal of emojis and links, • Removal of special characters, punctuation and numbers, • Removal of stop words, • Lemmatization. For simplicity, a bit vector was chosen to represent each single document. Subsequently, the cleaned dataset was divided into samples in terms of 5-fold cross-validation before HBBO could be applied to each sample, resulting in an optimized feature set for each fold. In order to observe the specific behavior of the new algorithm, the problem is considered as several binary classification tasks. In this respect, for each class, a model is trained using the training data without applying any kind of balancing technique. Table 1 show the values chosen for the experiments. Table 1 Summary of hyperparameter. Phase of algorithm Hyperparameter Value Initialization Number of different data sets for cross validation 5 Number of documents (individuals) per subset during optimization 24 Number of iterations for all phases 125 Education Number of terms for exchange 5 Percent of documents for adaptation of features 100 Consultation Number of terms for exchange 3 Percent of documents for adaption of features 100 Field changing Number of changed labels 1 Figure 2 shows the history of the 𝐹1 -score for all 125 iterations for the category true. Shown in green color is the best group of individuals in each iteration, while black dots show the mean of 𝐹1 -score of all groups and the red horizontal line symbolizes the baseline 𝐹1 -score trained without HBBO feature selection. The results of the remaining classes are shown in Figure 3, Figure 4, and Figure 5 accordingly. In order to provide a quantitative overview, the respective results are summarized in Table 2. In addition, the mean value is calculated for each class. The system yields a 𝑚𝑎𝑐𝑟𝑜 𝐹 1 = 0.602 over all classes. Table 2 Results in terms of 𝐹1 -Score for each fold of a 5-fold cross-validation. 𝐹1 -Score Class Fold 1 Fold 2 Fold 3 Fold 4 Fold 5 Mean True 0.508 0.464 0.553 0.558 0.574 0.531 False 0.765 0.762 0.777 0.787 0.764 0.771 Partially False 0.633 0.601 0.633 0.602 0.645 0.623 Other 0.476 0.45 0.468 0.416 0.607 0.483 As mentioned earlier, the results of all folds must be merged to combine them into one classifier. This can be achieved simply by concatenating the best group of individuals in each fold. The final feature matrix contains 120 pseudo documents (5 groups of 24 features each). In order to uniquely assign each document of the test data to a class, each separately optimized binary model was used for classification in the order of their performance. Table 3 summarizes the results of this final classification procedure. Here, 𝑚𝑎𝑐𝑟𝑜 𝐹1 = 0.251 with an accuracy of 0.462. With this result we have reached the 18𝑡ℎ place of the shared task 3a. Figure 2: HBBO results after 125 iterations for first category: true; black horizontal line: own baseline without HBBO, black: 𝐹1 -score of the best group of individuals, grey: mean 𝐹1 -score of all groups of individuals. Table 3 Evaluation results using hold-out data. Class Precision Recall 𝐹1 -Score True 0.391 0.086 0.141 False 0.573 0.806 0.670 Partially False 0.161 0.179 0.169 Other 0.016 0.032 0.022 6. Discussion The results shown are surprisingly low. Nevertheless, HBBO as a feature selection algorithm has a high potential for classification tasks. However, there are still some problems and open research questions. First of all, the drop of 𝑚𝑎𝑐𝑟𝑜 𝐹1 from training (0.602) and test (0.251) data, Figure 3: HBBO results after 125 iterations for second category: false; black horizontal line: own baseline without HBBO, black: 𝐹1 -score of the best group of individuals, grey: mean 𝐹1 -score of all groups of individuals. which might indicate a slight overfitting. The optimization of each sample takes into account the respective test set. This process may lead to good classification results only within that particular test set. Further studies need to be performed to test this hypothesis. Further performance gains could be achieved by structured experimentation with the hyper- parameters or by using more advanced features, such as TF-IDF or BM25. Inseparable from the chaining of binary classifiers is the question of their best order. Among other strategies, the shift to true multi-label classification could also be beneficial. For this, further adaptations of the HBBO algorithm have to be made. In addition, the strict requirement for improvement at each optimization step can lead to miss optimal feature combinations. This could be remedied by allowing temporary deteriorations. In general, the task of automatically carrying out a fine-grained classification of a document with regard to its truthfulness is difficult to imagine. Inseparable from the spread of fake news is the task of making it appear as real as possible. Thus, there can be no statistically detectable linguistic features that clearly indicate the truthfulness of a document, as can be Figure 4: HBBO results after 125 iterations for third category: partially false; black horizontal line: own baseline without HBBO, black: 𝐹1 -score of the best group of individuals, grey: mean 𝐹1 -score of all groups of individuals. easily demonstrated analytically. Ultimately, this can only be determined by a fact check. This insight is also underlined by the overall results achieved by all participants in this shared task. 7. Conclusion In this paper a novel feature selection algorithm for text classification using an human behavior- based optimization approach is presented in order to solve task of fine-grained fake news detection. The algorithm shows an improved performance compared to classification using a single SVM. In addition, the enormous reduction in training input after optimization is remarkable. In each sample, only 24 optimized pseudo-documents were able to outperform the baseline calculated considering all documents. Nevertheless, further experiments must be carried out to find the best values for the hyperpa- rameters. Further improvements might achieved by using a more advanced term representation. Figure 5: HBBO results after 125 iterations for fourth category: other; black horizontal line: own baseline without HBBO, black: 𝐹1 -score of the best group of individuals, grey: mean 𝐹1 -score of all groups of individuals. References [1] P. Nakov, A. Barrón-Cedeño, G. Da San Martino, F. Alam, J. M. Struß, T. Mandl, R. Míguez, T. Caselli, M. Kutlu, W. Zaghouani, C. Li, S. Shaar, G. K. Shahi, H. Mubarak, A. Nikolov, N. Babulkov, Y. S. Kartal, J. Beltrán, The clef-2022 checkthat! lab on fighting the covid-19 infodemic and fake news detection, in: M. Hagen, S. Verberne, C. Macdonald, C. Seifert, K. Balog, K. Nørvåg, V. Setty (Eds.), Advances in Information Retrieval, Springer Interna- tional Publishing, Cham, 2022, pp. 416–428. [2] P. Nakov, A. Barrón-Cedeño, G. Da San Martino, F. Alam, J. M. Struß, T. Mandl, R. Míguez, T. Caselli, M. Kutlu, W. Zaghouani, C. Li, S. Shaar, G. K. Shahi, H. Mubarak, A. Nikolov, N. Babulkov, Y. S. Kartal, J. Beltrán, M. Wiegand, M. Siegel, J. Köhler, Overview of the CLEF-2022 CheckThat! lab on fighting the COVID-19 infodemic and fake news detection, in: Proceedings of the 13th International Conference of the CLEF Association: Information Access Evaluation meets Multilinguality, Multimodality, and Visualization, CLEF ’2022, Bologna, Italy, 2022. [3] J. Köhler, G. K. Shahi, J. M. Struß, M. Wiegand, M. Siegel, T. Mandl, Overview of the CLEF-2022 CheckThat! lab task 3 on fake news detection, in: Working Notes of CLEF 2022—Conference and Labs of the Evaluation Forum, CLEF ’2022, Bologna, Italy, 2022. [4] S.-A. Ahmadi, Human behavior-based optimization: A novel metaheuristic approach to solve complex optimization problems, Neural Comput. Appl. 28 (2017) 233–244. URL: https://doi.org/10.1007/s00521-016-2334-4. doi:10.1007/s00521-016-2334-4. [5] G. K. Shahi, D. Nandini, FakeCovid – a multilingual cross-domain fact check news dataset for covid-19, in: Workshop Proceedings of the 14th International AAAI Conference on Web and Social Media, 2020. URL: http://workshop-proceedings.icwsm.org/pdf/2020_14.pdf. [6] G. K. Shahi, J. M. Struß, T. Mandl, Overview of the clef-2021 checkthat! lab task 3 on fake news detection, Working Notes of CLEF (2021). [7] G. K. Shahi, Amused: An annotation framework of multi-modal social media data, arXiv preprint arXiv:2010.00502 (2020). [8] G. K. Shahi, A. Dirkson, T. A. Majchrzak, An exploratory study of covid-19 misinformation on twitter, Online Social Networks and Media 22 (2021) 100104. [9] M. Sharma, P. Kaur, A comprehensive analysis of nature-inspired meta-heuristic techniques for feature selection problem, Archives of Computational Methods in Engineering 28 (2020). doi:10.1007/s11831-020-09412-6. [10] J. Xi, M. Spranger, D. Labudde, Music event detection leveraging feature selection based on ant colony optimization 13 (2020) 36–47. [11] R. Soto, B. Crawford, F. G. Molina, R. Olivares, Human behaviour based optimization supported with self-organizing maps for solving the s-box design problem, IEEE Access 9 (2021) 84605–84618. doi:10.1109/ACCESS.2021.3087139. [12] A. Wagdy, A. Hadi, A. Khater, Gaining-sharing knowledge based algorithm for solving optimization problems: a novel nature-inspired algorithm, International Journal of Machine Learning and Cybernetics 11 (2020). doi:10.1007/s13042-019-01053-x. [13] M. Kumar, A. J. Kulkarni, S. C. Satapathy, Socio evolution & learning optimization algo- rithm: A socio-inspired optimization methodology, Future Generation Computer Systems 81 (2018) 252–272. doi:https://doi.org/10.1016/j.future.2017.10.052.