<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An Oppositional Thinking Analysis Method Using BERT-based Model with BiGRU</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Qingbiao Hu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhongyuan Han</string-name>
          <email>hanzhongyuan@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jiangao Peng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mingcan Guo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chang Liu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Foshan University</institution>
          ,
          <addr-line>Foshan</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>The Oppositional thinking analysis: Conspiracy theories vs critical thinking narratives task of PAN at CLEF 2024 involves two challenges: first, distinguishing between conspiracy and critical narratives as Subtask 1, and second, identifying key elements of oppositional narratives as Subtask 2. We consider these two challenges as binary classification and sequence labeling problems, respectively. We will perform both tasks in English and Spanish. In this paper, we introduce our method to address these challenges by fine-tuning a BERT-based model with an added BiGRU layer for Subtask 1 and employing a multi-task learning method for Subtask 2. Finally, our model for English achieves MCC scores of 0.821 in Subtask 1 and Span-F1 scores of 0.569 in Subtask 2 on the oficial test set.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2024</kwd>
        <kwd>Oppositional Thinking Analysis</kwd>
        <kwd>BERT-based Model</kwd>
        <kwd>Multi-task Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Oppositional thinking analysis Task</title>
      <p>At PAN 2024 there are two subtasks proposed for oppositional thinking analysis:
• Subtask 1: Distinguishing between critical and conspiracy texts. It is a binary classification
task that aims to distinguish between two types of messages: the first contains critical messages
that scrutinize significant decisions within the public health sector without endorsing a
conspiratorial mindset; the second includes messages that interpret the pandemic or public health
decisions as the result of a malignant conspiracy orchestrated by secretive, powerful entities. Our
task is to categorize these texts into distinct categories: CONSPIRACY or CRITICAL.
• Subtask 2: Detecting elements of the oppositional narratives. It is a token-level classification
task aimed at recognizing text spans corresponding to the key elements of oppositional narratives.
A span-level annotation scheme that identifies the Agents (A), Facilitators (F), Campaigners (C),
Victims (V), Efects (E), Objectives (O) in the oppositional narratives was developed. Our task is
to identify specific spans in texts that should be annotated with the corresponding labels.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Method</title>
      <p>Generally speaking, our method consists of two main parts: the BERT-based encoder and the BiGRU
downstream neural network layer for both Subtask 1 and Subtask 2. Our method involves three primary
steps: 1) fine-tune the pre-trained BERT-based model with the given training dataset, 2) feed the
sequence of embeddings from the BERT-based model into a BiGRU layer and 3) Use the outputs from
the BiGRU layer, typically the final hidden states that encapsulate the information from the entire
sequence, to classify the text into categories (e.g., critical or conspiracy) in Subtask 1 or to combine
with diferent task heads for span annotation in Subtask 2.</p>
      <sec id="sec-3-1">
        <title>3.1. BERT-based Model with BiGRU Layer Architecture for Subtask 1</title>
        <p>In this section, we introduce the architecture for Subtask 1. Figure 1 shows the whole architecture.</p>
        <p>CRITICAL or
CONSPIRACY</p>
        <p>Softmax
Linear Layer
Dropout Layer</p>
        <p>Concatenated</p>
        <p>Hidden States</p>
        <p>BiGRU Layer
BERT-based Encoder</p>
        <p>Output
Sequence Output</p>
        <p>Tokenization
⋯⋯</p>
        <p>Input Data
word 1
word N</p>
        <p>The CT-BERT model is selected as our encoder, which was trained on a large dataset of COVID-19
Twitter messages. The corpus for this PAN 2024 task consists of COVID-19 Telegram texts, making
our model particularly well-suited due to its training on similar content. Consequently, this model is
expected to outperform other BERT-based models due to its superior understanding of this specific
domain. Additionally, we have chosen RoBERTa [9] as a contrasting model to verify whether these
expectations hold.</p>
        <p>The BERT-based model provides rich contextual embeddings by considering the left and right contexts
within the transformer architecture. The addition of a BiGRU layer introduces an extra level of sequential
processing. It processes information in both forward and backward directions across the text, ofering
a comprehensive view of the temporal dependencies. Once the BERT-based layer has generated the
sequence outputs, they are fed into the BiGRU layer. The BiGRU layer synthesizes the information
captured by the BERT layer, adding a layer of understanding. This enhancement aids in detecting subtle
cues and patterns that diferentiate various narrative types.</p>
        <p>The BiGRU outputs are then passed through additional dropout layers for regularization, followed by
a linear classification layer that maps the BiGRU outputs to the target category.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Multi-task Learning Architecture for Subtask 2</title>
        <p>The core architecture for Subtask 2 remains the same, however, we employ a multi-task learning method
to more efectively address the specific challenges posed by Subtask 2, as shown in Figure 2.</p>
        <p>Category:O
Start char:2
End char:135</p>
        <p>Token
Classification
+BiGRU</p>
        <p>Output
BIO Tagging</p>
        <p>Task Modules
For Different</p>
        <p>Categories
BERT-based Encoder</p>
        <p>Shared Layer</p>
        <p>Input Text</p>
        <p>Given that the key elements to be identified in a text fall under one of six categories— Agents (A),
Facilitators (F), Campaigners (C), Victims (V), Efects (E), and Objectives (O)—each can be considered
a separate token classification task. All these tasks share the same need for embeddings. Therefore,
we utilize a BERT-based encoder (primarily CT-BERT) as the backbone of our architecture, with token
classification layers serving as task-specific heads. This forms our multi-task classifier architecture.
Additionally, the token classification layer is integrated with a BiGRU layer, and through BIO tagging,
we achieve the span output for each category.</p>
        <p>Recent research [10] has proven the efectiveness of a multi-task classifier based on the domain-specific
CT-BERT model. Utilizing a shared encoder, our model eficiently learns universal representations
beneficial across all tasks, while the dedicated task modules concentrate on task-specific features.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <sec id="sec-4-1">
        <title>4.1. Datasets</title>
        <p>Given these two subtasks, the oppositional thinking analysis task has provided datasets [11] consisting
of Telegram texts related to COVID-19 from a list of oppositional Telegram channels, available in both
English and Spanish. The data has been pre-processed and tokenized for convenience, with emojis
and other non-text content removed. The training datasets include lists of texts fully annotated with
categories and spans of key elements, whereas the test datasets contain only the input texts. A total of
5000 texts for each language have been provided.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Evaluation</title>
        <p>For evaluation, we used the oficial metrics provided to evaluate Subtask 1: Matthews Correlation
Coeficient (MCC) [12], per-class F1 scores: F1-Consp and F1-Crit and macro-averaged F1.</p>
        <p>And we used the following metrics in Subtask 2: span-F1 [13], span-recall, span-precision and
micro-span-F1.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Baseline</title>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Settings</title>
        <p>The organisers of each subtask provided baselines in both languages for each subtask. BERT classifier
is used for Subtask 1, and BERT-based multi-task token classifier is used for Subtask 2.
While training, we preprocessed the training set and divided it using stratified 3-fold cross-validation.</p>
        <p>Our model is trained using a cross-entropy loss function and utilizes the AdamW optimizer with a
learning rate of 2e-5, incorporating a scheduler for learning rate adjustments. Other hyperparameters
include a batch size of 16 and a training duration of three epochs.</p>
        <p>In Subtask 1, we selected CT-BERT and RoBERTa for experiments on the English corpus, and
bertspanish [14] for the Spanish corpus. Each model was tested both with and without an added BiGRU
layer. In Subtask 2, we selected CT-BERT as backbone on the English corpus, and bert-spanish for the
Spanish corpus. Each model was tested both with an added BiGRU layer.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Results</title>
        <p>During the training process for Subtask 1, we evaluated our models and compared them with the
oficial baselines. We anticipate that the CT-BERT + BiGRU model will outperform other models on
the English corpus. For the Spanish corpus, due to the limited availability of multilingual models for
experimentation, we used BERT-Spanish with a BiGRU layer.</p>
        <p>As shown in Table 1, our model performed better than both the baseline and RoBERTa + BiGRU,
demonstrating the efectiveness of the CT-BERT + BiGRU model in this binary classification task. When
compared with CT-BERT without the BiGRU, the version with BiGRU showed slight improvement.
However, the BERT-Spanish + BiGRU model slightly fell short of the Spanish baseline.</p>
        <p>The Table 2 shows that our model still holds up, indicating that our model is robust and neither
overfits nor underfits the training set. However, the BERT-Spanish + BiGRU model performed worse
than the baseline.</p>
        <p>In relation to Subtask 2, and similar to the approach in Subtask 1, we compared the CT-BERT + BiGRU
model and the BERT-Spanish + BiGRU model with the baseline model during training to evaluate if this
multi-task architecture still performs better. Subsequently, we submitted our best model for testing on
the oficial test sets. Table 3 and Table 4 demonstrate the results obtained in Subtask 2.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This paper mainly introduces our work on oppositional thinking analysis at PAN 2024. Our work
utilizes a BERT-based model with a BiGRU layer to enhance performance in both binary classification
and sequence labeling tasks within this domain. The results from the oficial testing datasets indicate
that our method achieved an improvement of approximately 0.04 MCC scores in Subtask 1 and reached
4th place in the Oficial Ranking for the English corpus.</p>
      <p>While the English model demonstrated strong performance, the Spanish model was less successful,
with only marginal improvements attributed to the BiGRU layer. Therefore, future work should focus
on investigating how this method impacts multilingual tasks.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work is supported by the Social Science Foundation of Guangdong Province, China (No.GD24CZY02)
F. Rangel, N. Rizwan, P. Rosso, F. Schneider, A. Smirnova, E. Stamatatos, E. Stakovskii, B. Stein,
M. Taulé, D. Ustalov, X. Wang, M. Wiegmann, S. M. Yimam, E. Zangerle, Overview of PAN 2024:
Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking
Analysis, and Generative AI Authorship Verification, in: L. Goeuriot, P. Mulhem, G. Quénot,
D. Schwab, L. Soulier, G. M. D. Nunzio, P. Galuščáková, A. G. S. de Herrera, G. Faggioli, N. Ferro
(Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of
the Fifteenth International Conference of the CLEF Association (CLEF 2024), Lecture Notes in
Computer Science, Springer, Berlin Heidelberg New York, 2024.
[4] D. Korenčić, B. Chulvi, X. B. Casals, M. Taulé, P. Rosso, F. Rangel, Overview of the Oppositional
Thinking Analysis PAN Task at CLEF 2024, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S.
de Herrera (Eds.), Working Notes of CLEF 2024 – Conference and Labs of the Evaluation Forum,
2024.
[5] K. Pogorelov, D. T. Schroeder, S. Brenner, J. Langguth, FakeNews: Corona Virus and Conspiracies</p>
      <p>Multimedia Analysis Task at MediaEval 2021., in: MediaEval, 2021.
[6] J. Alghamdi, Y. Lin, S. Luo, Towards covid-19 fake news detection using transformer-based models,</p>
      <p>Knowledge-Based Systems 274 (2023) 110642.
[7] M. Müller, M. Salathé, P. E. Kummervold, Covid-twitter-bert: A natural language processing model
to analyse covid-19 content on twitter, Frontiers in artificial intelligence 6 (2023) 1023281.
[8] J. Chung, C. Gulcehre, K. Cho, Y. Bengio, Empirical evaluation of gated recurrent neural networks
on sequence modeling, arXiv preprint arXiv:1412.3555 (2014).
[9] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, V. Stoyanov,</p>
      <p>Roberta: A robustly optimized bert pretraining approach, arXiv preprint arXiv:1907.11692 (2019).
[10] Y. Peskine, G. Alfarano, I. Harrando, P. Papotti, R. Troncy, Detecting COVID-19-Related Conspiracy</p>
      <p>Theories in Tweets., in: MediaEval, 2021.
[11] D. Korenčić, B. Chulvi, X. Bonet Casals, M. Taulé, P. Rosso, PAN24 Oppositional Thinking Analysis
[Data set], https://doi.org/10.5281/zenodo.11199642, 2024. Available from Zenodo.
[12] D. Chicco, N. Tötsch, G. Jurman, The Matthews correlation coeficient (MCC) is more reliable
than balanced accuracy, bookmaker informedness, and markedness in two-class confusion matrix
evaluation, BioData mining 14 (2021) 1–22.
[13] G. Da San Martino, Y. Seunghak, A. Barrón-Cedeno, R. Petrov, P. Nakov, et al., Fine-grained analysis
of propaganda in news article, in: Proceedings of the 2019 conference on empirical methods
in natural language processing and the 9th international joint conference on natural language
processing (EMNLP-IJCNLP), Association for Computational Linguistics, 2019, pp. 5636–5646.
[14] J. Cañete, G. Chaperon, R. Fuentes, J.-H. Ho, H. Kang, J. Pérez, Spanish pre-trained bert model and
evaluation data, arXiv preprint arXiv:2308.02976 (2023).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>K. M. Douglas</surname>
            ,
            <given-names>J. E.</given-names>
          </string-name>
          <string-name>
            <surname>Uscinski</surname>
            ,
            <given-names>R. M.</given-names>
          </string-name>
          <string-name>
            <surname>Sutton</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Cichocka</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Nefes</surname>
            ,
            <given-names>C. S.</given-names>
          </string-name>
          <string-name>
            <surname>Ang</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Deravi</surname>
          </string-name>
          , Understanding conspiracy theories,
          <source>Political psychology 40</source>
          (
          <year>2019</year>
          )
          <fpage>3</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Phadke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Samory</surname>
          </string-name>
          , T. Mitra,
          <article-title>What makes people join conspiracy communities? Role of social factors in conspiracy engagement</article-title>
          ,
          <source>Proceedings of the ACM on Human-Computer Interaction</source>
          <volume>4</volume>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ayele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. B.</given-names>
            <surname>Casals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Elnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Freitag</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Korenčić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moskovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          , M. Potthast,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>