<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>FoSIL at CheckThat! 2022: Using Human Behaviour-Based Optimization for Text Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andy Ludwig</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jenny Felser</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jian Xi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dirk Labudde</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Spranger</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Applied Sciences Mittweida</institution>
          ,
          <addr-line>Technikumplatz 17, 09664 Mittweida</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Nowadays, a huge amount of information and news articles are available every day. The events of recent years have shown that Fake News can severely shake trust in politics and science. Unfortunately, a decision can only be made about the truthfulness of a fraction of all news and posts. In this respect, the CLEF2022-CheckThat! shared task 3a adresses this problem. In this paper, we propose a new classification approach using a novel metaheuristic feature selection algorithm that mimics human behavior. The results show that the performance of a baseline classifier can achieve higher performance by combining with this algorithm with only a fraction of the features.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;fake news detection</kwd>
        <kwd>text classification</kwd>
        <kwd>feature selection</kwd>
        <kwd>human behavior-based optimization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In times of constant availability of vast amounts of information, people have to judge the truth
of news in a short time. This assessment is often neglected due to the fast-moving nature of
the news. There are various reasons why authors, whether intentionally or unintentionally,
contribute to the generation of untrustworthy content. In particular, sources that deliberately
disseminate false information pose a danger to consumers of news.</p>
      <p>Efective methods for detecting fake news are essential in the fight against the targeted spread
of fake news. Continuous research and development of approaches to detect misinformation
are highly important. This is one mission of the CLEF2022 - CheckThat! Lab [1, 2]. In general,
the Lab’s goal is to verify the veracity of claims. Task 3a takes up the challenge of assessing the
truth content of news articles [3].</p>
      <p>This paper presents a novel approach for text classification. The concept is based on human
behaviour-based optimization (HBBO) [4]. This meta heuristic optimization approach uses
some fundamental interactions and behaviours of humans. The potential of this adaptation was
used for the fake news detection in task 3a.</p>
      <p>The paper is structured as follows: section 2 presents related works, section 3 describes the
human behaviour-based optimization, section 4 summarizes the adaption of the optimization
approach for text classification, section 5 presents the given data and the conducted experiments,
in section 6 the results are discussed and finally in section 7 a conclusion is given.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>The assessment of the truthfulness of news or claims is a special case of text classification whose
importance has increased greatly in recent times. Especially during the Covid-19 pandemic
situation a larger group of people was challenged to judge news correctly. This situation makes
the invention of automatic text classification systems necessary [ 5]. To pool a lot of competence
and to push the development forward, a task for diferent participants can be done like the
annual task CheckThat! Lab since 2018. The results of last year are summarized in [6]. These
tasks need a high quality dataset [7], which is also reflected in the available labels [8].</p>
      <p>To improve the results for fake news detection and the related text classification task, a wide
range of possible approaches can be explored. In this paper, the research focus was on a new
technique to select the best features for training. To solve these optimization tasks, nature
inspired meta-heuristic techniques can be applied. An overview of already known approaches
is given in [9]. For example, the ant colony optimization is a possible algorithm for feature
selection in text classification tasks [10].</p>
      <p>The nature-inspired algorithm used in this paper is derived from human behaviour. The basis
of this approach was presented in [4]. This algorithm was used already in [11] in combination
with self-organizing maps in context of cryptanalytic resistance. The authors compare this
approach with other meta-heuristic solutions like ant colony optimization. Next to the approach
of [4], other algorithms based on human behaviour were described. Firstly, in [12] the goal is to
solve optimization tasks by simulating the phases of knowledge gaining and sharing in younger
and older years of human life. A diferent aspect of optimization algorithms under usage of
human behaviour is presented in [13]. This approach focuses on the adaption of behaviours
and manners of other humans for example in family structures.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Human Behaviour-Based Optimization (HBBO)</title>
      <p>In this section, we briefly describe Ahmadi’s novel swarm intelligence-based optimization
approach [4] considering human behavior, which forms the basis for the feature selection
approach used here. A central aspect in all phases of the algorithm is an individual’s pursuit of
self-optimization at diferent stages of his or her life. At the same time, diferent individuals
have diferent levels of experience in their field, and some of them become experts in one of
them (e.g., art, music, or science, etc.).</p>
      <p>
        The optimization of individual performance is carried out iteratively in four main steps:
Initialization, Education, Consultation, and Field Changing. In the Initialization step the population
is built. Each individual in the population is assigned an area of interest in which
improvement should be achieved. Depending on the underlying optimization problem, an individual
is represented as a vector of characteristic variables (features).  = (︀ 1 2 . . .  )︀  . An
expert is determined for each field. The expert provides the best function value depending on
his individual feature set. The formal definition of an expert is shown in equation (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), where 
denotes a person,  denotes an expert, and  denotes the actual field of expertise.
( ) → ( ) = argmin  error( )
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
The first step in the improvement process, is Education. Individuals in each area learn from
the respective expert. The learning process comprises the improvement of the individual
characteristic values by those of the expert and aims at the reduction of their own error function
value.
      </p>
      <p>A similar procedure is used in Consultation. In this stage, individuals can learn from any
other individual, not only from an expert. For this purpose, some variables are merged between
two randomly selected individuals. The consultation is called efective and the merged set of
variables is kept if the updated variables lead to better function values. Otherwise, the update is
reversed.</p>
      <p>The last step is Field Changing. Whether an individual changes his field of interest is calculated
using a rank probability method and the roulette wheel selection.</p>
      <p>After Initialization, the three remaining steps are repeated until a stop criterion is met, e.g.
if the average function values do not change (or change too little) or the number of iterations
reaches a maximum.</p>
      <p>This algorithm can be adapted to find an optimal feature set for text classification. The next
section explains the adaptation in detail.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Adaptation for Text Classification</title>
      <p>In order to adapt the HBBO approach for feature selection, the optimization objective, i.e., an
objective function, must be determined. Here, the 1-score is chosen on a test set. For this
purpose, the preprocessed dataset must be divided into training and test data. To obtain reliable
results and avoid overfitting, the data set is split in terms of k-fold cross-validation. Each fold
results in an optimal feature set in the form of a document term matrix (DTM) that is able to
classify the respective test data with the greatest possible success. Finally, the results of all the
folds are merged into a DTM that contains only the most successful features to classify the
entire dataset. For assessing the performance after each optimization step, a support vector
machine (SVM) is used.</p>
      <sec id="sec-4-1">
        <title>4.1. Initialization</title>
        <p>During Initialization, each input document is considered as an single individual and encoded in
the form of a bit vector. This vector represents an individual’s knowledge. Each individual is
assigned the class of the respective document, which also represents the field of interest. All
vectorized documents together form the entire population. In contrast to the original algorithm
where each individual is optimized, here subsets of individuals are optimized together. This
leads to a smaller set of documents achieving good or even better classification results than
the original set of all documents. A group of individuals contains all fields, which leads to a
simultaneous optimization of all classes within a group.</p>
        <p>For this purpose, the individuals are grouped into subsets, with each subset equally populated
with individuals of each class (stratified approach). Each group of individuals is optimized
separately. The size of the subsets is a hyperparameter.</p>
        <p>The remaining three steps are applied iteratively to each group of each fold. The number of
iterations needs to be specified as a hyperparameter.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Education</title>
        <p>
          In the Education step, a group of individuals must first be determined for each field that achieves
the lowest error with respect to the objective function on the test set according to equation (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ).
Each of these groups is considered to be the expert group for that particular field.
( ) → ( ) = argmax 1
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
Subsequently, a subset  of features of the expert group, considering all classes, is merged
with the features of other individuals, belonging to the remaining groups. This procedure
corresponds to the non-experts approaching the expert group by updating their feature vector.
The number of terms transferred in this step is another hyperparameter, as well as the number
of adapted individuals, i.e., documents. Equation (3) shows the formal definition of this step,
where ( ) denotes the ℎ individual  in the specific field  and  ⊆ ( ).
        </p>
        <p>( ) = {︃(( )),∪ , ioftheerrrowr(ise( ) ∪ ) &lt;  error(( )) (3)
The feature update is considered successful and the merged feature set is retained only
if the individual achieves an improvement with respect to the objective function, i.e., better
classification results. Otherwise, the previous feature set remains unchanged.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Consultation</title>
        <p>In Consultation, the basic procedure is very similar to that in the Education step. The main
diference is in the group of individuals from which the features for merging are taken. Regardless
of expert status, features are merged between two randomly selected groups of individuals.
This leads to greater heterogeneity of terms across all groups. Equation (4) shows the formal
definition of Consultation, where  is a subset of all features of individual  and  ⊆ ( ).</p>
        <p>( ) = {︃(( )),∪  , ioftheerrrowr(ise( ) ∪  ) &lt;  error(( )) (4)</p>
        <p>As in Education, the updated feature set is retained only if the 1-score improves; otherwise,
the previous terms remain unchanged. Again, the number of terms exchanged and the number
of individuals paired can be controlled using hyperparameters.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Field changing</title>
        <p>The last step - Field Changing - does not manipulate the individuals’ terms, but changes the
ifeld associated with them. The number of randomly selected individuals changing the field is
another hyperparameter.</p>
        <p>In a multilevel classification, an individual can simply change their area of interest to that
of the most successful expert group. For this purpose, all expert groups are ordered by their
1-score, as shown in equation (5), where () is the expert group in the respective area.
 = {︁(1), ..., () |  error (︁ (1) &lt; (2) &lt; ... &lt; ()︁)}︁
(5)
Afterwards, the expert group with the highest rank in  determines the field of the individual
willing to switch, as shown in equation (6).</p>
        <p>* = {︁|() = argmin()∈   (︁ ()︁)}︁
(6)
In the special case of a binary classification, the process simply reduces to a field change, i.e., an
inversion of the class label. Again, the new field is retained only if it leads to an improvement
in the objective function value.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments and Results</title>
      <p>In this work, the adaptation of HBBO, as discussed in section 4, was applied to the problem of
detecting fake news in the context of the CLEF2022-CheckThat! shared task 3a. The documents
provided were news articles in English, which were to be grouped into the classes true, false,
partially false and other with regard to potentially containing fake news. For training the model,
the provided training and development sets were combined so that the total training corpus
comprised 1, 264 documents, each of which was assigned to exactly one of the four mentioned
classes. The resulting class distribution of the training corpus is shown in Figure 1.</p>
      <sec id="sec-5-1">
        <title>False</title>
        <p>578
358
211
117</p>
      </sec>
      <sec id="sec-5-2">
        <title>True</title>
      </sec>
      <sec id="sec-5-3">
        <title>Other</title>
      </sec>
      <sec id="sec-5-4">
        <title>Partially False</title>
        <p>Before the HBBO algorithm can be applied, the input data must be cleaned. To this end, a
wide range of cleaning steps were performed:
• Combination of article and headline into one pseudo document,
• Replacement of newlines and tabs with white space,
• Removal of emojis and links,
• Removal of special characters, punctuation and numbers,
• Removal of stop words,
• Lemmatization.</p>
        <p>For simplicity, a bit vector was chosen to represent each single document. Subsequently, the
cleaned dataset was divided into samples in terms of 5-fold cross-validation before HBBO could
be applied to each sample, resulting in an optimized feature set for each fold.</p>
        <p>In order to observe the specific behavior of the new algorithm, the problem is considered
as several binary classification tasks. In this respect, for each class, a model is trained using
the training data without applying any kind of balancing technique. Table 1 show the values
chosen for the experiments.</p>
        <p>As mentioned earlier, the results of all folds must be merged to combine them into one
classifier. This can be achieved simply by concatenating the best group of individuals in each
fold. The final feature matrix contains 120 pseudo documents (5 groups of 24 features each).</p>
        <p>In order to uniquely assign each document of the test data to a class, each separately optimized
binary model was used for classification in the order of their performance. Table 3 summarizes
the results of this final classification procedure. Here,  1 = 0.251 with an accuracy of
0.462. With this result we have reached the 18ℎ place of the shared task 3a.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>Precision
The results shown are surprisingly low. Nevertheless, HBBO as a feature selection algorithm
has a high potential for classification tasks. However, there are still some problems and open
research questions. First of all, the drop of  1 from training (0.602) and test (0.251) data,
which might indicate a slight overfitting. The optimization of each sample takes into account
the respective test set. This process may lead to good classification results only within that
particular test set. Further studies need to be performed to test this hypothesis.</p>
      <p>Further performance gains could be achieved by structured experimentation with the
hyperparameters or by using more advanced features, such as TF-IDF or BM25. Inseparable from the
chaining of binary classifiers is the question of their best order. Among other strategies, the
shift to true multi-label classification could also be beneficial. For this, further adaptations of
the HBBO algorithm have to be made.</p>
      <p>In addition, the strict requirement for improvement at each optimization step can lead to miss
optimal feature combinations. This could be remedied by allowing temporary deteriorations.</p>
      <p>In general, the task of automatically carrying out a fine-grained classification of a document
with regard to its truthfulness is dificult to imagine. Inseparable from the spread of fake
news is the task of making it appear as real as possible. Thus, there can be no statistically
detectable linguistic features that clearly indicate the truthfulness of a document, as can be
easily demonstrated analytically. Ultimately, this can only be determined by a fact check. This
insight is also underlined by the overall results achieved by all participants in this shared task.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>In this paper a novel feature selection algorithm for text classification using an human
behaviorbased optimization approach is presented in order to solve task of fine-grained fake news
detection. The algorithm shows an improved performance compared to classification using
a single SVM. In addition, the enormous reduction in training input after optimization is
remarkable. In each sample, only 24 optimized pseudo-documents were able to outperform the
baseline calculated considering all documents.</p>
      <p>Nevertheless, further experiments must be carried out to find the best values for the
hyperparameters. Further improvements might achieved by using a more advanced term representation.
Bologna, Italy, 2022.
[3] J. Köhler, G. K. Shahi, J. M. Struß, M. Wiegand, M. Siegel, T. Mandl, Overview of the
CLEF-2022 CheckThat! lab task 3 on fake news detection, in: Working Notes of CLEF
2022—Conference and Labs of the Evaluation Forum, CLEF ’2022, Bologna, Italy, 2022.
[4] S.-A. Ahmadi, Human behavior-based optimization: A novel metaheuristic approach to
solve complex optimization problems, Neural Comput. Appl. 28 (2017) 233–244. URL:
https://doi.org/10.1007/s00521-016-2334-4. doi:10.1007/s00521-016-2334-4.
[5] G. K. Shahi, D. Nandini, FakeCovid – a multilingual cross-domain fact check news dataset
for covid-19, in: Workshop Proceedings of the 14th International AAAI Conference on Web
and Social Media, 2020. URL: http://workshop-proceedings.icwsm.org/pdf/2020_14.pdf.
[6] G. K. Shahi, J. M. Struß, T. Mandl, Overview of the clef-2021 checkthat! lab task 3 on fake
news detection, Working Notes of CLEF (2021).
[7] G. K. Shahi, Amused: An annotation framework of multi-modal social media data, arXiv
preprint arXiv:2010.00502 (2020).
[8] G. K. Shahi, A. Dirkson, T. A. Majchrzak, An exploratory study of covid-19 misinformation
on twitter, Online Social Networks and Media 22 (2021) 100104.
[9] M. Sharma, P. Kaur, A comprehensive analysis of nature-inspired meta-heuristic techniques
for feature selection problem, Archives of Computational Methods in Engineering 28
(2020). doi:10.1007/s11831-020-09412-6.
[10] J. Xi, M. Spranger, D. Labudde, Music event detection leveraging feature selection based
on ant colony optimization 13 (2020) 36–47.
[11] R. Soto, B. Crawford, F. G. Molina, R. Olivares, Human behaviour based optimization
supported with self-organizing maps for solving the s-box design problem, IEEE Access 9
(2021) 84605–84618. doi:10.1109/ACCESS.2021.3087139.
[12] A. Wagdy, A. Hadi, A. Khater, Gaining-sharing knowledge based algorithm for solving
optimization problems: a novel nature-inspired algorithm, International Journal of Machine
Learning and Cybernetics 11 (2020). doi:10.1007/s13042-019-01053-x.
[13] M. Kumar, A. J. Kulkarni, S. C. Satapathy, Socio evolution &amp; learning optimization
algorithm: A socio-inspired optimization methodology, Future Generation Computer Systems
81 (2018) 252–272. doi:https://doi.org/10.1016/j.future.2017.10.052.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Beltrán,</surname>
          </string-name>
          <article-title>The clef-2022 checkthat! lab on fighting the covid-19 infodemic and fake news detection</article-title>
          , in: M.
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Verberne</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Seifert</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Nørvåg</surname>
          </string-name>
          , V. Setty (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer International Publishing, Cham,
          <year>2022</year>
          , pp.
          <fpage>416</fpage>
          -
          <lpage>428</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Míguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babulkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Kartal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beltrán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Köhler</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2022 CheckThat! lab on fighting the COVID-19 infodemic and fake news detection</article-title>
          ,
          <source>in: Proceedings of the 13th International Conference of the CLEF Association: Information Access Evaluation meets Multilinguality</source>
          , Multimodality, and Visualization, CLEF '
          <year>2022</year>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>