<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>JU_NLP at CheckThat! 2025: A Confidence-guided Transformer-based Approach for Multilingual Subjectivity Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Srijani Debnath</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dipankar Das</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science &amp; Engineering, Government College of Engineering and Leather Technology</institution>
          ,
          <addr-line>Kolkata</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science &amp; Engineering, Jadavpur University</institution>
          ,
          <addr-line>Kolkata</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>With the rapid progress of transformer-based multilingual models, accurately distinguishing subjective authorial viewpoints from objective reportage in news articles has become increasingly challenging. This paper presents a comprehensive framework for binary subjectivity classification that combines fine-tuned multilingual BERT with a lightweight, feature-based post-processing module for English. We fine-tuned bert-base-multilingual-cased on five languages-Arabic, Bulgarian, English, German, and Italian-under monolingual, multilingual, and zero-shot conditions, and evaluated on held-out data in Greek, Polish, Ukrainian, and Romanian. For the English monolingual subtask, low-confidence predictions (confidence&lt;0.7) were refined using two lexical features: a SentiWordNet-based subjective score and a combined lexical-clues plus opinionlexicon count. Our approach achieved macro-averaged F1 scores of 0.545 for Arabic (rank-14), 0.733 for English (rank-8), 0.699 for Italian (rank-10), 0.736 for German (rank-9), 0.435 for Greek (rank-8), 0.560 for Polish (rank-11), 0.580 for Ukrainian (rank-11), 0.744 for Romanian (rank-8), and 0.654 in the multilingual setting (rank-10).</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;subjectivity classification</kwd>
        <kwd>multilingual BERT</kwd>
        <kwd>post-processing</kwd>
        <kwd>macro F1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Early approaches to subjectivity and sentiment classification relied on shallow machine learning models
with hand-crafted features. Pang and Lee (2004) employed n-gram features with Support Vector
Machines to achieve strong performance on English sentiment data [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Lexicon-based methods, such
as those using subjectivity lexicons by Wiebe et al. (2002), extracted polarity scores and worked well in
monolingual settings but struggled to generalize across domains and languages [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        With the advent of neural networks, convolutional and recurrent architectures became popular. Kim
(2014) demonstrated that a simple CNN over word embeddings could outperform traditional
featurebased models on various text classification tasks [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. BiLSTM models with attention further improved
recall of subtle subjective cues, but these methods required large amounts of annotated data in each
target language.
      </p>
      <p>
        Transformer-based pretrained models revolutionized cross-lingual transfer. Devlin et al. (2019)
introduced BERT, and its multilingual variant (mBERT) showed strong zero-shot performance across
over 100 languages. Subsequent work by Barnes et al. (2020) specifically fine-tuned mBERT for
crosslingual subjectivity detection, demonstrating consistent gains over BiLSTM baselines [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Despite these advances, subjectivity classification in low-resource and unseen languages remains
challenging. Our work builds on the strengths of mBERT’s cross-lingual embeddings and introduces a
lightweight post-processing module for English to refine low-confidence predictions using interpretable
lexical features.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>
        All experiments in this study were conducted using the annotated dataset provided by the CheckThat!
2025 Lab: Task 1 – Subjectivity Detection [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13">10, 11, 12, 13</xref>
        ]. The dataset is multilingual and spans five
primary training languages—English, Italian, German, Bulgarian, and Arabic—with additional zero-shot
evaluation on five other languages, namely Ukrainian, Romanian, Greek, Polish, and a multilingual mix.
Each training language includes data splits for training, development, and development-test, while the
evaluation phase was performed on an unseen test set. The dataset consisted of sentences extracted
from news articles in multiple languages which were annotated as either subjective (SUBJ) or objective
(OBJ) as per the guidelines developed by Antici et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] that defines a sentence to be subjective if its
content is based on or influenced by personal feelings, tastes, or opinions. Otherwise, the sentence is
objective. Table 1 presents the dataset statistics for each of the five primary languages. In general, the
dataset exhibits class imbalance where number of objective sentences is more than that of subjective
ones across majority of the languages.
      </p>
      <p>In addition to Table 1, the English test set contained 300 sentences, the Italian set had 299, the German
test set included 347 sentences, and Arabic had the largest test set with 1036 senetnces. Notably, no
oficial test data was released for Bulgarian.</p>
      <p>Beyond the core training languages, the test set also included data for evaluating zero-shot
performance. Specifically, the Ukrainian test set contained 297 sentences, the Romanian set had 206, the
Greek test set included 284, and Polish comprised 351 test instances. Furthermore, a multilingual test
set with 1982 sentences drawn from a diverse mix of languages was provided to assess cross-lingual
generalization capabilities.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>This section describes the methodologies we employed for subjectivity classification. Given an input
sentence  , our goal was to determine whether  expressed a subjective viewpoint (SUBJ) or an
objective statement (OBJ).</p>
      <sec id="sec-4-1">
        <title>4.1. Text Preprocessing</title>
        <p>All raw sentences were first cleaned and normalized. We removed escape characters such as ‘ \n’
(newline) and ‘\t’ (tab), and converted any Unicode characters to their ASCII equivalents. After
cleaning, each sentence was tokenized into a sequence of word tokens [1, 2, . . . , ] using the NLTK
tokenizer.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Textual Analysis and Lexical Feature Extraction</title>
        <p>For the English monolingual subtask, we defined two interpretable, lexicon-based features to refine
low-confidence BERT predictions.</p>
        <sec id="sec-4-2-1">
          <title>4.2.1. SentiWordNet Subjective Score</title>
          <p>Inspired by Esuli and Sebastiani, we computed a subjective score for each sentence based on
SentiWordNet synsets. Each token was POS-tagged and mapped to a WordNet tag; for each unique lemma,
we averaged the sum of positive and negative scores across its synsets. The sentence-level score was
obtained by summing these averages:</p>
          <p>ScoreSWN( ) =</p>
          <p>1
∑︁ ∑︁ (︀ pos_score() + neg_score())︀ ,
ℓ∈ℒ( ) |(ℓ)| ∈(ℓ)
where ℒ( ) is the set of lemmatized tokens in  and (ℓ) are the SentiWordNet synsets for lemma ℓ.
Sentences with ScoreSWN ≥ 1.615 were considered subjective.</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Combined Lexical-Clues and Opinion-Lexicon Count</title>
          <p>
            We counted occurrences of lexical clues that often reflect subjectivity, such as “perhaps”, “I believe”,
“wonderful”, “wow”, “my pleasure”, “sorry”, and others. Some of these phrases were adapted from the
human-phrase list proposed in [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ], which showed that such expressions are more commonly used by
humans than AI which can be linked with personal opinions and subjective content. Since our task
focused on identifying subjective language, these clues were suitable indicators.
          </p>
          <p>In addition, we included extra phrases that matched our task requirements — specifically hedge words
(e.g., “maybe”, “likely”, “suppose”) and sarcastic or emotional expressions (e.g., "yeah right", "totally", "as
if"), as these also signal subjectivity. All the lexical clues that we considered are provided in Appendix
A. To further strengthen this feature, we also counted sentiment-bearing words from the Hu and Liu
Opinion Lexicon. Let Φ be the set of human-phrases and Ω the set of opinion-lexicon words. For each
sentence  , we computed</p>
          <p>Count( ) =
∑︁ freq(,  ) + ∑︁ freq(,  ).
∈Φ
∈Ω
Sentences with Count( ) ≥ 4 were labeled as subjective.
and the final label was chosen as</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Training</title>
      <sec id="sec-5-1">
        <title>4.2.3. Threshold Selection Methodology</title>
        <p>To choose the best threshold for each feature, we experimented with several values of threshold by
performing binary classification with each value on the English training dataset and checked which one
gave the highest classification accuracy. For the SentiWordNet Subjective Score, we tried thresholds
such as 1.5, 1.6, 1.615, 1.625, 1.65, 1.7 etc. Among these, 1.615 gave the best accuracy in separating
subjective and objective sentences, so we used it as the final threshold.</p>
        <p>For the Combined Lexical-Clues and Opinion-Lexicon Count, we experimented with thresholds like
2, 3, 4, 5 etc. A threshold of 4 gave the best balance — it correctly identified most of the subjective
sentences without misclassifying too many objective ones.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Framework Development</title>
      <p>
        The overall framework is shown in Figure 1 where SentiW denotes the SentiWordNet Subjective Score
feature and the Feat denotes Combined Lexical-Clues and Opinion-Lexicon Count feature. We built our
classifier on top of the transformer-based bert-base-multilingual-cased (mBERT) model [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
We selected mBERT because it was pretrained on the Wikipedia dumps of 104 languages, yielding strong
contextual representations across diverse linguistic families, and because it has demonstrated robust
zero-shot transfer capabilities without target-language fine-tuning [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. This multilingual pretraining
allows mBERT to leverage shared subword vocabularies and syntactic patterns, making it particularly
suitable for our nine-subtask setting that spans both high- and low-resource languages. Each input
sentence was tokenized using the mBERT tokenizer to produce input_ids and attention_mask
tensors. These tensors were passed through the mBERT encoder, yielding a sequence of hidden states
{h1, h2, . . . , h}.
      </p>
      <p>We then selected the hidden state corresponding to the special [CLS] token, denoted h[CLS], as the
aggregate representation of the entire sentence.</p>
      <p>Classification: The vector h[CLS] was fed into the built-in classification head of mBERT, which
applies dropout (with the default probability of 0.1) and a linear layer to produce two logits (0, 1).
These logits were converted into class probabilities via the softmax function:
 =</p>
      <p>exp()
exp(0) + exp(1)</p>
      <p>,  ∈ {0, 1},
ˆ = arg max ,</p>
      <p>
        ∈{0,1}
where ˆ = 1 indicates a subjective sentence and ˆ = 0 indicates an objective sentence.
We used the training and development splits provided by the CheckThat! 2025 Task 1 organizers, further
stratifying each split into 90% for training and 10% for validation. We fine-tuned the model for two
epochs with a per-device batch size of eight, using standard cross-entropy loss function and AdamW
optimizer [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] with a learning rate of 2 × 10− 5. We configured the Trainer to evaluate at the end of
each epoch, log training progress every 100 steps, and save model checkpoints after each epoch. All
experiments used a fixed random seed of 42 for reproducibility.
      </p>
    </sec>
    <sec id="sec-7">
      <title>7. Post-processing</title>
      <p>After training, we applied our feature-based post-processing only to the English monolingual test
predictions with confidence less than 0.7 to recompute its label by evaluating the two lexical features
described in Section 4.2. Specifically, each sentence with maximum predicted probability max  &lt; 0.7
was assigned SUBJ only if both its SentiWordNet subjective score exceeded 1.615 and its combined
lexicalclues plus opinion-lexicon count exceeded 4; otherwise, it was labeled OBJ. This strategy leveraged
interpretable linguistic cues to correct uncertain model outputs.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Experimental Setup</title>
      <p>All the experiments were accomplished using Python libraries such as pandas, nltk, datasets and
scikit-learn. The transformer-based mBERT model was trained and evaluated using HuggingFace’s
transformers library with support from PyTorch. The development and execution of the model
pipeline, including training and inference, were carried out entirely in the Kaggle environment equipped
with NVIDIA Tesla T4 dual GPUs.</p>
      <p>For English-specific post-processing, we used additional Python modules such as nltk’s
sentiwordnet, wordnet, opinion_lexicon, and scipy’s softmax function to calculate lexical
features. To evaluate the performance of the proposed subjectivity classification framework, macro-F1
scores were calculated for the exact test data provided by the organizers of the CheckThat! 2025 Task-1.</p>
    </sec>
    <sec id="sec-9">
      <title>9. Results</title>
      <p>We evaluated the results for the nine subtasks under the CheckThat! 2025 Task 1 using the fine-tuned
mBERT model. The task comprises three evaluation settings: monolingual, multilingual, and zero-shot.
The monolingual subtasks include Arabic, English, German, Italian, and Bulgarian; the multilingual
setting involves joint training and testing on all five training languages; and the zero-shot setting tests
on unseen languages—Greek, Polish, Ukrainian, and Romanian—after training on the multilingual
corpus.</p>
      <p>The macro-averaged F1 scores and the team rank of the mBERT classifier for each subtask are
provided in Table 2.</p>
      <p>Among all subtasks, the best performance was achieved on Romanian in the zero-shot setting with a
macro-F1 of 0.744. In contrast, the lowest result was obtained for the Greek zero-shot setting (0.435),
indicating a possible gap in representation coverage for low-resource languages with limited overlap
in the multilingual vocabulary. Our team achieved a strong rank of 8 in this setting in Romanian,
while we matched the same rank in Greek despite its low score. Romanian’s strong result may be
attributed to its linguistic similarity with Italian and the presence of shared subword units in mBERT’s
vocabulary, allowing better transfer from the multilingual training set. On the other hand, Greek, being
morphologically rich and having limited lexical overlap with the training languages, likely sufered
from poor representation alignment in the mBERT embedding space.</p>
      <p>In case of German and Italian languages, the model performed robustly, with macro-F1 scores of 0.736
and 0.699 respectively therefore achieving a team rank of 9 and 10 respectively, likely due to suficient
training data and the presence of high-resource status in the original pretraining corpus of mBERT.</p>
      <p>English, despite being one of the primary pretraining languages for mBERT, showed slightly lower
performance (0.693) on test dataset whereas it performed slightly better on development-test dataset
(0.730). The reason for low performance on English can be possibly due to the inherent dificulty of
distinguishing subjective expressions from objective ones within its dataset. The rank achieved on the
test set after post-processing is discussed later in Table 4.</p>
      <p>Arabic, with a macro-F1 of 0.545, underperformed relative to other monolingual settings, which may
stem from its complex morphology, right-to-left structure, and limited overlap with other Latin-based
training languages used in the multilingual setup. The model ranked 14th in this subtask.</p>
      <p>The multilingual setting, which involved joint training and testing on all five training languages,
resulted in a macro-F1 score of 0.654 with a team rank of 10. While this value was slightly lower
than some monolingual track performances, it confirmed our fine-tuned mBERT’s ability to generalize
reasonably well when trained on mixed-language data, balancing performance across a broader linguistic
spectrum.</p>
      <p>Result after post-processing: To further enhance classification in the English monolingual setting,
we applied a post-processing step on top of the mBERT classifier. The post-processing was applied
selectively to predictions with confidence lower than a threshold value  , using lexical features described
earlier in Section 4.5.</p>
      <p>Table 3 presents the macro-F1 scores on the English development-test split across multiple confidence
scores such as 0.99,0.95,0.90,0.85,0.75,0.70 and 0.65.</p>
      <p>From Table 3, it is evident that the post-processing technique consistently improved the macro-F1
scores compared to the baseline mBERT model, which scored a macro-F1 of 0.693 without any post-hoc
adjustments.</p>
      <p>Thresholds below conf &lt; 0.65 or above conf &lt; 0.75 resulted in diminishing macro-F1 gains, indicating
that the post-processing mechanism is only efective within this intermediate confidence range. Based
on the analysis of Table 3, the best performance was achieved at a confidence threshold of conf &lt;
0.70, where the macro-F1 peaked at 0.7575 due to which the value conf &lt; 0.70 was selected as the
optimal operating point. This threshold ofered a balanced trade-of between overall performance and
robustness against uncertain predictions.</p>
      <p>Using this chosen threshold, we applied the post-processing pipeline on the English subtask test
set to assess its efectiveness beyond the development data. The results before and after applying the
post-processing technique along with the team rank in this subtask are summarized in Table 4.</p>
      <p>As shown in Table 4, the macro-F1 score improved from 0.693 to 0.733 on the English test data after
post-processing achieving a remarkable team rank of 8 among all other teams. This result reinforces the
benefit of applying interpretable, lexical-based heuristics in combination with transformer predictions,
particularly when dealing with borderline or ambiguous classifications.</p>
      <p>Overall, the proposed mBERT-based subjectivity classification framework demonstrated promising
performance across diverse linguistic settings and was further strengthened by lightweight
postprocessing enhancements that addressed model uncertainty in a principled and explainable manner.
10. Conclusion
We presented a compact mBERT-based framework for multilingual subjectivity classification and
evaluated it across monolingual, multilingual, and zero-shot settings in CheckThat! 2025 Task 1. The
approach delivered strong performance in high-resource languages—German (0.736 macro-F1 score),
Italian (0.699 macro-F1 score) and English (0.693 macro-F1 score)—and achieved notable zero-shot
transfer for Romanian (0.744 macro-F1 score), while revealing challenges for morphologically rich
languages such as Greek and Arabic. The multilingual model attained a macro-F1 score of 0.654,
demonstrating reasonable cross-language generalization when trained on mixed-language data. A
lightweight post-processing step further boosted English performance, improving macro-F1 score
from 0.693 to 0.733 on the test set. These results demonstrated the value of combining transformer
representations with interpretable lexical heuristics for robust, explainable subjectivity detection across
diverse languages.
11. Limitations
Despite encouraging results, our work presents several limitations. First, performance varied notably
across languages, especially in zero-shot settings. The poor result on Greek suggests that mBERT’s
pretraining does not equally benefit all languages, particularly those with distinct orthographies or
limited token overlap with the training set. This calls for future research into adaptive multilingual
pretraining or low-resource augmentation strategies to close the performance gap.</p>
      <p>Second, the post-processing module was designed specifically for English, relying on lexical cues and
syntactic constructs that may not directly transfer to other languages. While the module improved both
overall and subjective-class accuracy, extending it to non-English settings would require developing
language-specific rules that respect local morphological and syntactic properties.</p>
      <p>Lastly, the classification was conducted at the sentence level without considering discourse or
contextual information. Integrating document-level context or inter-sentence dependencies could
help in resolving subtler subjectivity distinctions that span across sentences. Future work could also
explore hybrid architectures that blend LLMs with graph-based discourse models for deeper semantic
understanding.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT (OpenAI) for the sole purpose of
Paraphrase and reword. After using this tool, the authors reviewed and edited the content as needed
and take full responsibility for the publication’s content.</p>
      <p>Tools and services: ChatGPT (OpenAI)
Tools’ contributions (GenAI Usage Taxonomy): Paraphrase and reword.</p>
      <p>No generative AI system is listed as an author, and all core scientific contributions (problem formulation,
experiments, analyses, and conclusions) were performed solely by the human authors.</p>
    </sec>
    <sec id="sec-11">
      <title>A Appendix: Lexical-Clues</title>
      <p>The following list of Lexical Clues was used in our post-processing heuristic to help distinguish subjective
text. These phrases include sarcastic words, hedging expressions, personal references, and emotional
cues commonly found in subjective content.</p>
      <p>1. “therefore”
2. “however”
3. “etc”
4. “whatever”
5. “?”
6. “maybe”</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding (</article-title>
          <year>2019</year>
          )
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          , G. Lample,
          <article-title>Unsupervised cross-lingual representation learning at scale</article-title>
          , in: ACL,
          <year>2019</year>
          , pp.
          <fpage>8440</fpage>
          -
          <lpage>8451</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          , CoRR abs/
          <year>1810</year>
          .04805 (
          <year>2018</year>
          ). URL: http://arxiv.org/abs/
          <year>1810</year>
          .04805. arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Esuli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sebastiani</surname>
          </string-name>
          ,
          <article-title>Sentiwordnet: A publicly available lexical resource for opinion mining</article-title>
          ,
          <source>in: LREC</source>
          ,
          <year>2006</year>
          , pp.
          <fpage>417</fpage>
          -
          <lpage>422</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Mining and summarizing customer reviews</article-title>
          ,
          <source>in: Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '04,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2004</year>
          , p.
          <fpage>168</fpage>
          -
          <lpage>177</lpage>
          . URL: https://doi.org/10.1145/ 1014052.1014073. doi:
          <volume>10</volume>
          .1145/1014052.1014073.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts</article-title>
          ,
          <source>in: Proceedings of ACL</source>
          ,
          <year>2004</year>
          , pp.
          <fpage>271</fpage>
          -
          <lpage>278</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bruce</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <article-title>O'Hara, Recognizing subjective sentences: A subjectivity lexicon approach</article-title>
          ,
          <source>in: Proceedings of the ACL Workshop on Text Classification</source>
          ,
          <year>2002</year>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>52</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>Convolutional neural networks for sentence classification</article-title>
          ,
          <source>in: Proceedings of EMNLP</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1746</fpage>
          -
          <lpage>1751</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Barnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Cross-lingual subjectivity detection with multilingual bert</article-title>
          ,
          <source>in: Proceedings of EMNLP</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1234</fpage>
          -
          <lpage>1245</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. V.</surname>
          </string-name>
          ,
          <article-title>The clef-2025 checkthat! lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: C.
          <string-name>
            <surname>Hauf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Kazai</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nardini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pinelli</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Silvestri</surname>
          </string-name>
          , N. Tonellotto (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2025</year>
          , pp.
          <fpage>467</fpage>
          -
          <lpage>478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nawrocka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ivasiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Razvan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mihail</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! lab task 1 on subjectivity in news article</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.), Working Notes of CLEF 2025 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2025</year>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. V.</surname>
          </string-name>
          ,
          <article-title>The clef-2025 checkthat! lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: C.
          <string-name>
            <surname>Hauf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Kazai</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nardini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pinelli</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Silvestri</surname>
          </string-name>
          , N. Tonellotto (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2025</year>
          , pp.
          <fpage>467</fpage>
          -
          <lpage>478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Venktesh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! Lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: J.
          <string-name>
            <surname>Carrillo-de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Antici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bardi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fedotova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <article-title>A corpus for sentence-level subjectivity detection on English news articles</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            , M.-
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Kan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sakti</surname>
          </string-name>
          , N. Xue (Eds.),
          <source>Proceedings of the 2024 Joint International Conference on Computational Linguistics</source>
          ,
          <article-title>Language Resources and Evaluation (LREC-COLING 2024), ELRA</article-title>
          and
          <string-name>
            <given-names>ICCL</given-names>
            ,
            <surname>Torino</surname>
          </string-name>
          , Italia,
          <year>2024</year>
          , pp.
          <fpage>273</fpage>
          -
          <lpage>285</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .lrec-main.
          <volume>25</volume>
          /.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>U.</given-names>
            <surname>Jawaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Debnath</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. Das</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bandyopadhyay</surname>
          </string-name>
          ,
          <article-title>Human vs machine: An automated machine-generated text detection approach</article-title>
          , in: S. Lalitha Devi,
          <string-name>
            <surname>K.</surname>
          </string-name>
          Arora (Eds.),
          <source>Proceedings of the 21st International Conference on Natural Language Processing (ICON)</source>
          ,
          <article-title>NLP Association of India (NLPAI), AU-KBC Research Centre</article-title>
          , Chennai, India,
          <year>2024</year>
          , pp.
          <fpage>215</fpage>
          -
          <lpage>223</lpage>
          . URL: https://aclanthology. org/
          <year>2024</year>
          .icon-
          <volume>1</volume>
          .24/.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T.</given-names>
            <surname>Pires</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Schlinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Garrette</surname>
          </string-name>
          ,
          <article-title>How multilingual is multilingual bert?</article-title>
          ,
          <source>Proceedings of NAACLHLT</source>
          (
          <year>2019</year>
          )
          <fpage>4996</fpage>
          -
          <lpage>5001</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>I.</given-names>
            <surname>Loshchilov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hutter</surname>
          </string-name>
          ,
          <article-title>Decoupled weight decay regularization</article-title>
          ,
          <source>in: International Conference on Learning Representations</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>