<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Binary Battle: Leveraging Machine Learning and Transfer Learning Models to Distinguish between Conspiracy Theories and Critical Thinking</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sidharth Mahesh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sonith Divakaran</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kavya Girish</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hosahalli Lakshmaiah Shashirekha</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore, Karnataka</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>74</volume>
      <fpage>199</fpage>
      <lpage>209</lpage>
      <abstract>
        <p>In the context of automatic content moderation Natural Language Processing (NLP) has a complex task when it comes to distinguishing between conspiracy theories and critical thinking. While conspiracy theories present complex narratives attributing significant events to covert actions by powerful and malicious entities, critical thinking involves scrutinizing decisions without resorting to any sinister explanations. Making this distinction is essential to avoid the mislabeling of valid criticism as conspiracy, which may unintentionally lead people to join conspiracy communities. Conspiratorial and critical narratives are both examples of oppositional thinking, which is important in public debate, particularly in controversial areas like public health. In this direction, “Oppositional thinking analysis: Conspiracy theories vs critical thinking narratives"- a shared task organized at PAN 2024, invites the research community to address the challenges of distinguishing between conspiracy and critical texts in English and Spanish languages. To explore the strategies for distinguishing between critical and conspiracy texts in English and Spanish on social media platforms, in this paper, we - team MUCS, describe the models proposed for Subtask-1: "Distinguishing between critical and conspiracy texts" of the shared task. We explored machine learning models trained with Term Frequency-Inverse Document Frequency (TF-IDF) of char n-grams in the range (1, 5) and transfer learning techniques using several BERT variants fine-tuned with the given English and Spanish datasets, to classify the given unlabeled English and Spanish text into one of the two categories - 'CONSPIRACY' or 'CRITICAL'. Among the proposed models, English_BERT and Spanish_BERT models obtained Matthews Correlation Coeficient (MCC) scores of 0.7162 and 0.6293 for English and Spanish languages respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Oppositional thinking analysis</kwd>
        <kwd>Conspiracy vs Critical Narratives</kwd>
        <kwd>Oppositional Thinking</kwd>
        <kwd>Conspiracy Theories</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Transfer Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Conspiracy theories and critical thinking are two forms of oppositional thinking which are common
especially on contentious issues and analyzing oppositional thinking entails scrutinizing narratives that
question mainstream perspectives. Further, understanding the impact of diferent types of oppositional
thinking on public opinion and behavior is crucial [1]. While critical thinking fosters constructive
and democratic debate characterized by reasoned questioning without unfounded explanations,
conspiratorial thinking can lead to misinformation and social conflict attributing the significant events to
hidden malevolent forces[2]. In social and political discourse, conspiracy theories can have detrimental
efects on an individual as well as on organisations or the entire society. These theories suggest that
major social (encompass gatherings and activities involving people) and political events (activities
related to government and leadership) with careful planning by powerful and malevolent entities can
spread false information and stir social unrest. Conspiracy theories have been associated with violence,
war, terrorism, prejudice, poor health choices, and denial of climate change [3]. In contrast, critical
thinking involves analyzing and questioning decisions, particularly in areas like public health, without
attributing events to hidden conspiracies.</p>
      <p>Mislabeling critical discourse as conspiratorial can suppress healthy debate and alienate individuals
who are merely questioning decisions and mislabeling critical conversation as conspiracy can stifle
constructive disagreement and alienate people who are only challenging judgements [3, 4]. This leads
to a climate of mistrust and inhibits the exchange of open and sincere ideas. On the other hand, failing
to identify and address conspiratorial narratives allows the spread of misinformation, leading to societal
division and mistrust. Thus, accurate diferentiation between the two forms of oppositional thinking
by content moderation systems is essential to avoid marginalizing legitimate criticism and preventing
individuals from being drawn into conspiracy communities by mistaking valid critique for conspiracy
theories [5]. Additionally, distinguishing between these forms of oppositional thinking is essential for
public discourse and social harmony.</p>
      <p>
        Distinguishing between the oppositional forms of thinking is challenging due to the nuanced and
overlapping content, the context-dependent nature of oppositional statements, and oppositional attitudes
[6]. Further, diferentiating between well-founded critical and unfounded conspiracy theories require
advanced and context-aware NLP techniques and developing such techniques is crucial for improving
the accuracy and fairness of content moderation systems. To address the challenges of distinguishing
between the oppositional forms of thinking, "Oppositional thinking analysis: Conspiracy theories vs
Critical thinking narratives" shared task organized at PAN 2024 [7], invites the research community to
develop models to distinguish between conspiracy and critical thinking texts in English and Spanish
languages. The shared task has two subtasks in English and Spanish languages and we - team MUCS
participated in only Subtask-1: ’Distinguishing between critical and conspiracy texts’. This subtask
having two categories - ’CONSPIRACY’ and ’CRITICAL’, is modeled as a binary classification problem.
In this paper, we describe various machine learning models trained with TF-IDF of char n-grams in the
range (
        <xref ref-type="bibr" rid="ref1">1, 5</xref>
        ) and transfer learning techniques using several BERT variants fine-tuned with the given
Engish and Spanish datasets, to classify the given unlabeled English and Spanish text into one of the
two categories - ’CONSPIRACY’ or ’CRITICAL’. The sample text from the given datasets for English
and Spanish are shown in the Tables 1 and 2 respectively.
      </p>
      <p>The rest of the paper is organized as follows: Section 2 describes the recent literature on the two
forms of oppositional thinking and Section 3 focuses on the description of the proposed models followed
by the experiments and results in Section 4. The paper concludes with future works in Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Conspiracy theories involve elaborate, unverified claims driven by cognitive biases and emotional
needs, often rejecting oficial explanations without suficient evidence. In contrast, critical thinking is
characterized by objective analysis, logical reasoning, and the evaluation of evidence from multiple
sources. Research highlights that while conspiracy beliefs are linked to cognitive biases and feelings of
powerlessness, critical thinking fosters informed decision-making and intellectual humility [3].</p>
      <p>To explore diferent strategies for the conspiracy textual content classification in social media,
Moosleitner and Murauer [8] employed machine learning models (Support Vector Machines (SVM),
Multinomial Naive Bayes (MNB), and Extremely randomized Trees) trained with TF-IDF of character,
word and Document Term n-grams based features and BERT models (BERT-base, RoBERTa, and
DistilBERT) for English text. For the three tasks: Task 1 - Text-Based Misinformation Detection, Task 2
- Text-Based Conspiracy Theories Recognition, and Task 3 - Text-Based Combined Misinformation and
Conspiracies Detection, their proposed BERT-base models outperformed all other models in all three
tasks obtaining MCC scores of 0.3184, 0.3624, 0.3347 for Tasks 1, 2, and 3 respectively. For detecting fake
news during covid-19 pandemic, Tahat et al. [9] proposed a hybrid analysis using Structural Equation
Modelling (SEM) and machine learning classification algorithms such as BayesNet, AdaBoostM1, LWL,
Logistic, J48, and OneR, for English dataset. Among the proposed models J48 classifier outperformed
other machine learning classifiers with an F-Measure score of 0.863. Peskine et al. [10] proposed
transformer model (composed of an ensembling of CT-BERT models) and a node embedding-based
techniques (node2vec + Multilayer Perceptron (MLP) classification head) to detect COVID-19-related
conspiracy theories in tweets in English language which consists of three subtasks: Task 1 - Text-Based
Misinformation and Conspiracies Detection, Task 2 - Graph-Based Conspiracy Source Detection, and
Task 3 - Graph and Text-Based Conspiracy Detection. Their proposed CT-BERT ensembling model
obtained a MCC score of 0.710 and 0.719 for Task 1 and Task 3 respectively, and node2vec + MLP model
obtained a MCC of 0.355 for Task 2.</p>
      <p>Giachanou et al. [2] performed a comparative analysis of various profiles and psychological and
linguistic characteristics in social media texts of users who share posts about conspiracy theories. The
authors then compared the efectiveness of these characteristics for predicting whether a user is a
conspiracy propagator or not by proposing ConspiDetector, a model that is based on a Convolutional
Neural Network (CNN) which combines word embeddings with psycho-linguistic characteristics
extracted from the tweets of users to detect conspiracy propagators. Recordare et al. [11] implemented
various machine learning classifiers (Logistic Regression (LR), k-Nearest Neighbours (kNN), Naive
Bayes (NB), SVM, Decision Trees (DT), Random Forest (RF), Gradient Boosting: XGBoost and LightGBM,
Quadratic Discriminant Analysis, MLP, Ridge Classifier, and Linear Discriminant Analysis, trained
with Bidirectional Auto-Regressive Transformers (bart)-large-Multi-Genre natural language inference
features for identifying users who propagate conspiracy theories based on a rich set of 871 features in
English language. Among all the proposed models, LightGBM classifier outperformed other models
with a macro F1 score of 0.87. To identify whether an article belongs to conspiracy theory or not in
English language, Ghasemizade and Onaolapo [12] proposed machine learning classifiers (RF, SVM,
k-NN, NB) trained with TF-IDF of word unigrams and deep learning model trained with padded and
embedded text sequences. Using their respective tokenizers, tokenized and padded text sequences were
taken as inputs to train the transformer models - BERT and RoBERTa. Their proposed RoBERTa model
outperformed other models with a macro F1 score of 87%.</p>
      <p>The above literature highlights extensive research eforts aimed at detecting conspiracy theories
utilizing a range of machine learning, deep learning, and transfer learning models. These studies ofer
valuable insights into the detection of conspiracy theories. However, they do not specifically address
the distinction between conspiracy theories and critical thinking within the framework of oppositional
thinking. This gap suggests a need for further research to efectively distinguish between these concepts,
encouraging the creation of new models for this specific type of application.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>We have explored machine learning and transfer learning models for distinguishing between critical
and conspiracy texts in English and Spanish and the steps involved in the construction of these models
are explained in the following subsections.</p>
      <sec id="sec-3-1">
        <title>3.1. Machine Learning models</title>
        <p>The framework of machine learning model is visualized in Figure 1 and the steps included in building
these classifiers are explained below:</p>
        <sec id="sec-3-1-1">
          <title>3.1.1. Pre-processing</title>
          <p>Pre-processing is the preliminary step in building learning models and it involves cleaning and
transforming the raw text data to a suitable format required for subsequent processing. Usually, text data
contains noise in the form of: user mentions, hashtags, punctuation, digits, and hyperlinks, and
eliminating this irrelevant information makes the data less complex and improves the performance of the
classifier. Hence, in this work, this irrelevant information and stopwords are removed during
preprocessing. Further, English and Spanish stopwords available at NLTK library1 are used as references
to remove English and Spanish stopwords respectively from the given dataset.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.1.2. Feature Extraction</title>
          <p>
            The role of feature extraction is to extract relevant features from the given data to train the learning
models. TF-IDF of char n-grams is used to represent English and Spanish text. Char n-grams are
sequences of n consecutive characters in a word and char n-grams in the range (
            <xref ref-type="bibr" rid="ref1">1, 5</xref>
            ) are obtained from
the text and converted to TF-IDF vectors using Tfidf Vectorizer 2.
1https://www.nltk.org/search.html?q=stopwords
2https://scikit-learn.org/stable/modules/generated/sklearn.feature_extraction.text.TfidfVectorizer.html
          </p>
        </sec>
        <sec id="sec-3-1-3">
          <title>3.1.3. Model Construction</title>
          <p>The performance of the learning model relies on the features and the classifier used to carry out
classification. This work utilizes machine learning classifiers (MNB, LR, RF, and ensemble of LSVC,
LR, and RF with majority voting), to distinguish between conspiracy and critical texts in English and
Spanish language. A brief description of the machine learning classifiers is given below:
• MNB - is a probability-based classifier suitable for text classification with discrete characteristics
like word frequency counts [13].
• LinearSVC - used from the Scikit-learn library3 attempts to maximize the distance between
classified samples by finding a hyperplane.
• LR - is used to predict the probability of certain classes based on dependent variables and is
suitable for binary classification task. Further, regularisation approaches in LR classifiers are
useful for reducing overfitting in high dimensional space [14].
• RF - is one of the supervised learning algorithms which is flexible and can be adapted easily to
diferent situations but it is necessary to build a minimum number of trees in order to classify the
data [15].
• Ensemble learning - is a strategy for building a new classifier from several heterogeneous base
classifiers taking benefit of the strength of one classifier to overcome the weakness of another
classifier to get better performance for the classification task [ 16]. In this work, three machine
learning classifiers (LSVC, LR, and RF) are ensembled with hard voting to distinguish between
critical and conspiracy texts.</p>
          <p>The hyperparameters and their values used in the machine learning models are shown in Table 3.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Transfer Learning</title>
        <p>The technique of transfer learning within the broader field of machine learning utilizes knowledge
gained from one task to improve the performance of another related task. This is realized with the help
of pretrained transformer models which are trained on large unlabeled data and are widely accessible
and applicable to various tasks. The pretrained transformer models are fine-tuned with the given dataset
to fit the models to a particular task or domain. The framework of the proposed transfer learning model
is shown in Figure 2.</p>
        <p>The text is pre-processed to clean and transform the raw text into a consistent format by converting
numeric information to corresponding words, and removing URLs, user mentions, hash tags, and special
characters. Preprocessing is applied to the sentences of the given text to retain the sentence structure
in the text and the preprocessed text is used to fine-tune the transformer models. A brief description of
the transformer models used in this study to fine-tune are given below:
3https://scikit-learn.org/stable/modules/generated/sklearn.svm.LinearSVC.html
• BERT_base4 is a conceptually simple and empirically powerful pretrained language model using
a Masked Language Modeling (MLM) objective trained on Toronto Book Corpus and Wikipedia
and exclusively used for tasks involving English texts.
• English_BERT5 - is a bilingual Legal BERT model trained with 2,000 Dutch and 6,000 English
legal documents amounting to 12 GB legal text from various areas belonging to legal domain such
as legislation and court cases. This domain specific BERT has resulted in improved performance
compared to using standard BERT models for legal tasks.
• CT_BERT_v26 - is a BERT-large-uncased model pretrained on a corpus of messages from Twitter
about COVID-19. This model is identical to covid-twitter-bert but trained on more data (40.7M
sentences and 633M tokens) resulting in higher performances for many downstream applications.
• EN_RoBERTa7 - is a large multi-lingual language model trained on 12.5TB of filtered
CommonCrawl data. Based on Facebook’s RoBERTa model, this model is fine-tuned with the conll2003
dataset in English.
• ES_BERT8 - BETO: Spanish BERT is trained on the Spanish edition of Wikipedia, the OPUS
Project, and Spanish books and news articles, and is exclusively used for tasks involving Spanish
texts.
• Distil_SpanBERT9 - is a distilled version of SpanishBERT, trained on Spanish text sources, and is
also exclusively used for tasks involving Spanish texts, but is optimized for eficiency and speed.
• Spanish_BERT10 - is a sentence-transformers model which maps sentences and paragraphs to a
768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
• ES_RoBERTa11 - a variant of RoBERTa is a BERT-based model, specifically tailored for the
Spanish language. It is trained on large Spanish text corpora to understand and generate contextually
relevant representations of words and sentences.</p>
        <p>We employed the above mentioned BERT variants from the Hugging Face library. The
hyperparameter and their values used in the above transfer learning models are shown in Table 4. These
BERT variants are fine-tuned with the pre-processed Train set and is used to train transformer
classiifer (ClassificationModel) to distinguish between conspiracy and critical texts in English and Spanish
language.
4https://huggingface.co/google-bert/bert-base-uncased
5https://huggingface.co/Gerwin/legal-bert-dutch-english
6https://huggingface.co/digitalepidemiologylab/covid-twitter-bert-v2
7https://huggingface.co/FacebookAI/xlm-roberta-large-finetuned-conll03-english
8https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased
9https://huggingface.co/dccuchile/distilbert-base-spanish-uncased
10https://huggingface.co/hiiamsid/sentence_similarity_spanish_es
11https://huggingface.co/bertin-project/bertin-roberta-base-spanish</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <p>The datasets provided by the organizers of the shared task consisted of only Train set and are highly
imbalanced. Statistics of the datasets are as follows:
• English dataset: 2,621 samples belong to ’CRITICAL’ class and 1,379 samples belong to
’CON</p>
      <p>SPIRACY’ class.
• Spanish dataset: 2,538 samples belong to ’CRITICAL’ class and 1,462 samples belong to
’CON</p>
      <p>SPIRACY’ class.</p>
      <p>
        As the datasets consists of only Train sets, 33% of the Train sets at random are considered as Validation
sets to evaluate the performances of the proposed models for both the languages and the remaining
as the Train sets. Experiments were carried out by training various machine learning models using
TF-IDF of char n-grams in the range (
        <xref ref-type="bibr" rid="ref1">1, 5</xref>
        ) and by fine-tuning various BERT variants mentioned above,
to distinguish between conspiracy and critical texts in English and Spanish. The performances of the
proposed models were evaluated on the Validation set based on macro F1 score and the performances
are shown in Tables 5 and 6 for English and Spanish datasets respectively.
      </p>
      <p>The results shown in Tables 5 and 6 illustrate that transfer learning models have performed better
than machine learning models. As the shared task participants were allowed to submit the predictions
of only two models on the Test sets, we fine-tuned English_BERT and EN_RoBERTa with the complete
English Train set and Spanish _BERT and ES_RoBERTa with the complete Spanish Train set and the
predictions of these models on English and Spanish Test sets submitted to the organizers were evaluated
based on MCC scores. MCC is a metric used to evaluate the quality of binary classifications especially
with imbalanced datasets. The MCC value ranges from -1 to +1, where +1 indicates perfect prediction, 0
indicates no better than random prediction, and -1 indicates total disagreement between prediction and
observation. MCC provides a balanced and comprehensive measure of model’s performance, considering
all types of classification errors. MCC scores are calculated using the following formula:
(  ×   ) − (  ×   )</p>
      <p>MCC = √︀(  +   )(  +   )(  +   )(  +   )
where:
•   (True Positives) - number of correct positive predictions,
•   (True Negatives) - number of correct negative predictions,
•   (False Positives) - number of incorrect positive predictions,</p>
      <p>Among the two transfer learning models submitted to the shared task, English_BERT and
Spanish_BERT models obtained MCC scores of 0.7162 and 0.6293 for English and Spanish texts securing
61st and 36th position respectively. The performances of these models is shown in the Table 7. The
low performances of fine-tuned English_BERT and Spanish_BERT models could be attributed to the
following reasons:
• The given datasets are highly imbalanced with approximately 2/3 of the total samples belonging
to ’CRITICAL’ class and 1/3 belonging to ’CONSPIRACY’ class in both the languages. The highly
imbalanced data will significantly impact the models’ performance with a bias towards majority
class.
• English_BERT designed for both Dutch and English may potentially reduce its efectiveness for
purely English tasks. Further, domain-specific models like English_BERT might not generalize
well outside their specialized contexts.</p>
      <p>• Spanish_BERT model might be more suited for sentence similarity tasks rather than classification.
Further, diferences in pre-training data, fine-tuning processes, and the values of hyperparameters used
in the models could also contribute to the disparity in the performances of the proposed models.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>
        In this paper, we - team MUCS, describe the models submitted to Subtask-1: ’Distinguishing between
critical and conspiracy texts’ of the shared task "Oppositional thinking analysis: Conspiracy theories vs
critical thinking narratives" at ’PAN 2024’, to distinguishing between critical and conspiracy texts in
English and Spanish. Experiments are carried out with TF-IDF of char n-grams in the range (
        <xref ref-type="bibr" rid="ref1">1, 5</xref>
        ) to
train several machine learning classifiers and several BERT variants are fine-tuned with the English
and Spanish Train sets in transfer learning models, to distinguish the given unlabeled English and
Spanish text as ’CONSPIRACY’ or ’CRITICAL’. As the shared task participants were allowed to submit
the predictions of only two models on the Test set, we fine-tuned English_BERT and EN_RoBERTa
with the complete English Train set and Spanish_BERT and ES_RoBERTa with the complete Spanish
Train set and the predictions of these models on English and Spanish Test sets were submitted to the
organizers for evaluation. Among these two models, English_BERT and Spanish_BERT obtained MCC
scores of 0.7162 and 0.6293 for English and Spanish texts securing 61st and 36th position respectively.
As the given datasets are imbalanced, suitable text augmentation techniques followed by eficient text
representation methods and context-aware models to distinguish between the two forms of oppositional
thinking will be explored further.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Sharp</surname>
          </string-name>
          , Oppositional Ideas, Not Dichotomous Thinking: Reply to Rorty, in: Political Theory, Sage Publications Sage CA: Los Angeles, CA,
          <year>2010</year>
          , pp.
          <fpage>142</fpage>
          -
          <lpage>147</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Giachanou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ghanem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <article-title>Detection of Conspiracy Propagators using Psycho-Linguistic Characteristics</article-title>
          , in
          <source>: Journal of Information Science, SAGE Publications Sage UK: London, England</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>