<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>NapierNLP at CheckThat! 2025: Detecting Subjectivity with LLMs and Model Fusion</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Katarina Alexander</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Md Zia Ullah</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dimitra Gkatzia</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computing, Engineering and the Built Environment, Edinburgh Napier University</institution>
          ,
          <addr-line>Scotland</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents the participation of our team, NapierNLP, in the CLEF2025 CheckThat! Lab Task 1: Subjectivity. This task aimed to diferentiate between subjective and objective sentences within a corpus sourced from news articles. We formulate the task as a classification problem and fine-tune three pre-trained large language models (LLMs): GPT2, Qwen, and GPT-Neo. To enhance our predictions, we combine the outputs from multiple models using a majority voting method. We conducted experiments and evaluated our approach using the CheckThat! 2025 Task 1 dataset. Our results showed that GPT-Neo outperforms the other two models. In the oficial competition results, we ranked 18th in the monolingual English category, but our combination method proved to be more efective than the individual models we employed.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Subjectivity Detection</kwd>
        <kwd>Natural Language Processing (NLP)</kwd>
        <kwd>Large Language Models (LLMs)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Previous research on subjectivity detection primarily relied on rule-based templates and patterns
that were bounded by the datasets used, leading to a lack of generalizability [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The introduction of
deep-learning models enhanced this generalizability by utilising labelled data to learn models capable
of recognising key features of subjective sentences. Team HYBRINFOX, which won the CLEF 2024
CheckThat! Task 2 on subjectivity detection, leveraged a BERT-based approach, specifically RoBERTa
combined with VAGO, achieving a macro F1 score of 0.74 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Large Language Models (LLMs) can be efectively used for language-based classification tasks [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
including subjectivity detection [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Shokri et al. experimented with GPT-3.5, GPT-4, and Gemini on
three datasets: MPQA [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], News-1 [10], and News-2 [11]. They found that even a zero-shot approach
performed well at the task, with an average of 0.71, 0.71, and 0.69, respectively across the datasets. LLMs
were also used in the CheckThat! 2024 competition. Team SemanticCuetSync [12] used Llama-3-8b
and achieved a macro F1 score of 0.72 in the Arabic test set and 0.50 in the English test set. Meanwhile,
Team CLaC-2 [13] used Google’s Gemini, alongside prompt engineering, to achieve a macro F1 of 0.45
on the English test set.
      </p>
      <p>
        Using an ensemble model can yield better results than individual models [14]. Majority voting is
a straightforward technique for classification problems, where individual models classify sentences
independently, and the final prediction is based on the most frequently assigned label [ 15]. Shokri
et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] used this technique to enhance their model’s performance, achieving a macro F1 of 0.74, an
increase of 0.03 over their best individual model.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>3.1. LLMs
The task of detecting whether a sentence is subjective or objective can be framed as a binary classification
problem. By utilising human-annotated texts, a model can be developed using a supervised learning
technique to classify sentences accordingly.</p>
      <p>LLMs are large-scaled, pre-trained neural language models that have emergent abilities that are not
present in smaller models [16]. LLMs can then be fine-tuned for a variety of downstream tasks, including
sentence classification.</p>
      <p>In this study, we fine-tune three publicly available LLMs from HuggingFace: GPT2 [ 17], Qwen2.5 [18,
19], and GPT-Neo [20]. Details of the models can be found in Table 1.</p>
      <p>We anticipate that a fusion model could outperform individual models, so we employed majority
voting as our ensemble technique. Each model provides a prediction label, and the common label across
all three models is used as our final classification.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Evaluations</title>
      <p>In this section, we describe the datasets, evaluate our approach, and discuss the results.</p>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>The English language corpus consisted of sentences sourced from news articles [21], along with an ID
and a label: “OBJ" for objective and “SUBJ" for subjective. These had been annotated by humans based
on the following criteria:
“A sentence is subjective if its content is based on or influenced by personal feelings, tastes,
or opinions. Otherwise, the sentence is objective." [22]</p>
        <p>Table 2 provides a breakdown of all datasets. The training dataset and both testing datasets are
imbalanced towards objective sentences. This is expected given the data source, as news articles are
typically presented as an objective recounting of events. No augmentation was applied to the training
set, as initial experiments demonstrated that the model performed well despite the imbalance.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experimental settings</title>
        <p>We fine-tune the LLMs listed in Table 1 on the training data for 100 epochs using
AutoModelForSequenceClassification and evaluated using the development set. Due to size constraints, both Qwen and
GPT-Neo were trained using a parameter-eficient fine-tuning method called LoRA [ 23]. The training
was conducted on an NVIDIA GeForce RTX 2080 Super with 8 GB GPU memory.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Results and Discussions</title>
        <p>Performance on Development Test Set: Table 3 shows our results for each model, including the
ifnal combined model. The baseline was a logistic regressor trained on a multilingual SentenceBERT
model, as provided by competition organisers. In the development test data, GPT2 achieved higher F1
scores for both macro (0.75) and subjective (0.62) metrics. However, the combined model demonstrated
an equal accuracy of 0.81 and outperformed all other models for the objective class with an F1 of 0.88.
Performance on Test Set: Given GPT2’s performance on the development test data, we decided to
submit it as our final run for the competition. On the test data, GPT2 achieved an F1 of 0.67, placing
18th overall.</p>
        <p>The results of all four models on the test data can be seen in Table 5. Among the three individual
models, GPT-Neo performed the best, matching the combined model with a macro F1 score of 0.72.
Notably, unlike the results on the development test data, GPT-Neo outperformed GPT2 in the subjective
F1 score, achieving 0.61 compared to GPT2’s 0.56. Qwen performed the worst, scoring the lowest across
all tests. This lower performance is likely due to insuficient training time, as it was trained for 100
epochs, achieving a macro F1 of 0.92 on the training set, compared to F1 of 1.00 for both GPT2 and
GPT-Neo.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Failure Analysis</title>
        <sec id="sec-4-4-1">
          <title>Years later, in Korea, I was much happier, but I still OBJ</title>
          <p>wanted to drown out the reality of the $60,000 of
student loans I felt like I was spending a lifetime
paying of and my feelings of impostor syndrome.</p>
        </sec>
        <sec id="sec-4-4-2">
          <title>Despite the hardships, however, revolutionary op- OBJ</title>
          <p>timism was palpable.</p>
        </sec>
        <sec id="sec-4-4-3">
          <title>The chancellor asserts, but can’t possibly be confi- SUBJ</title>
          <p>dent, that planes will be landing on a third runway
by 2035.</p>
        </sec>
        <sec id="sec-4-4-4">
          <title>The reality is that Labour’s wider agenda does not SUBJ</title>
          <p>support the broad-based growth that would truly
produce those improvements.</p>
        </sec>
        <sec id="sec-4-4-5">
          <title>Speaking delegates condemned imperialists and OBJ</title>
          <p>warmongers, and shared tales of national struggles
and workplace victories.</p>
        </sec>
        <sec id="sec-4-4-6">
          <title>It must look elsewhere to make up for lost Russian OBJ</title>
          <p>gas.</p>
        </sec>
        <sec id="sec-4-4-7">
          <title>Planning delays, likewise, are the product of auster- OBJ</title>
          <p>ity and the resulting privatisation and outsourcing
of local government functions.</p>
        </sec>
        <sec id="sec-4-4-8">
          <title>If investors can be found to stump up that kind of OBJ</title>
          <p>cash, they will want a return.</p>
        </sec>
        <sec id="sec-4-4-9">
          <title>In a story that has been repeated countless times SUBJ</title>
          <p>across Europe, the right’s failed austerity agenda
has abetted the rise of the far right.</p>
        </sec>
        <sec id="sec-4-4-10">
          <title>So they become a diferent person. OBJ</title>
        </sec>
        <sec id="sec-4-4-11">
          <title>SUBJ</title>
        </sec>
        <sec id="sec-4-4-12">
          <title>SUBJ OBJ</title>
        </sec>
        <sec id="sec-4-4-13">
          <title>SUBJ</title>
          <p>OBJ
OBJ
OBJ</p>
        </sec>
        <sec id="sec-4-4-14">
          <title>SUBJ OBJ OBJ</title>
          <p>Each of the three individual models correctly identified some instances that the others failed to
classify accurately, ultimately contributing to incorrect final labels. Among the 72 errors in the final
combined model, 24 sentences were misclassified by all three models, with an equal distribution between
subjective and objective categories.</p>
          <p>For some sentences that were misclassified, the reasoning behind the errors can be identified. For
example, Sentence 3 in Table 6 was incorrectly predicted as objective by all models; however, the phrase
"but can’t possibly be confident" indicates that it is actually subjective. Without this clause, the sentence
could be correctly classified as objective. Similarly, sentences containing first-person pronouns, like
Sentence 1, may have been misclassified because this is a common feature of many subjective sentences.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>For our approach, we tested three diferent large language models and employed a majority voting
system to classify sentences as subjective or objective. Our results indicate that even with limited GPU
power, LLMs can efectively determine whether a sentence is subjective or objective. The majority
voting system matched or even outperformed the individual models, suggesting that combining model
outputs can yield more reliable predictions than relying on a single model.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Future Work</title>
      <p>There are several ways in which this research could be enhanced in future work. Using larger versions of
the models and allowing for longer training times could be feasible with improved GPU resources. This
GPT2
OBJ</p>
      <p>Qwen2.5</p>
      <p>GPT-Neo</p>
      <sec id="sec-6-1">
        <title>SUBJ</title>
      </sec>
      <sec id="sec-6-2">
        <title>SUBJ</title>
      </sec>
      <sec id="sec-6-3">
        <title>SUBJ OBJ</title>
      </sec>
      <sec id="sec-6-4">
        <title>SUBJ</title>
      </sec>
      <sec id="sec-6-5">
        <title>SUBJ</title>
        <p>OBJ
OBJ</p>
      </sec>
      <sec id="sec-6-6">
        <title>SUBJ OBJ OBJ</title>
      </sec>
      <sec id="sec-6-7">
        <title>SUBJ OBJ</title>
      </sec>
      <sec id="sec-6-8">
        <title>SUBJ</title>
        <p>OBJ</p>
      </sec>
      <sec id="sec-6-9">
        <title>SUBJ</title>
        <p>OBJ
OBJ</p>
      </sec>
      <sec id="sec-6-10">
        <title>SUBJ</title>
        <p>OBJ
OBJ
OBJ</p>
      </sec>
      <sec id="sec-6-11">
        <title>SUBJ</title>
        <p>OBJ
OBJ</p>
      </sec>
      <sec id="sec-6-12">
        <title>SUBJ</title>
      </sec>
      <sec id="sec-6-13">
        <title>SUBJ OBJ OBJ</title>
        <p>would also enable us to explore the use of other models, such as Llama [24] or Mistral [25]. Additionally,
we could investigate a prompting approach, either as an alternative to or in conjunction with our
current method.</p>
        <p>Another possible idea is to reevaluate how the final results are combined. Whilst majority voting
proved efective for this project, it could become problematic with an even number of models,
complicating the task of establishing a majority. Employing other ensemble methods, such as weighted voting
- where the models are assigned weighted on their results- or incorporating a confidence score for each
model could further improve our outcomes.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[10] F. Antici, A. Galassi, F. Ruggeri, K. Korre, A. Muti, A. Bardi, A. Fedotova, A. Barrón-Cedeño,
A corpus for sentence-level subjectivity detection on english news articles, 2024. URL: https:
//arxiv.org/abs/2305.18034. arXiv:2305.18034.
[11] E. Savinova, F. Moscoso Del Prado, Analyzing subjectivity using a transformer-based regressor
trained on naïve speakers’ judgements, in: J. Barnes, O. De Clercq, R. Klinger (Eds.), Proceedings
of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, &amp; Social Media
Analysis, Association for Computational Linguistics, Toronto, Canada, 2023, pp. 305–314. URL:
https://aclanthology.org/2023.wassa-1.27/. doi:10.18653/v1/2023.wassa-1.27.
[12] A. Paran, M. Hossain, S. Shohan, J. Hossain, S. Ahsan, M. Hoque, Semanticcuetsync at checkthat!
2024: Finding subjectivity in news articles using llama notebook for the checkthat! lab at clef 2024,
2024.
[13] S. Gruman, L. Kosseim, Clac-2 at checkthat! 2024: A zero-shot model for check-worthiness
and subjectivity classification, in: Conference and Labs of the Evaluation Forum, 2024. URL:
https://api.semanticscholar.org/CorpusID:271822050.
[14] Z.-H. Zhou, Ensemble Methods: Foundations and Algorithms, 1st ed., Chapman &amp; Hall/CRC, 2012.
[15] Fusion of Label Outputs, John Wiley &amp; Sons, Ltd, 2004, pp. 111–149. doi:https://doi.org/10.</p>
      <p>1002/0471660264.ch4.
[16] J. Wei, Y. Tay, R. Bommasani, C. Rafel, B. Zoph, S. Borgeaud, D. Yogatama, M. Bosma, D. Zhou,
D. Metzler, E. H. Chi, T. Hashimoto, O. Vinyals, P. Liang, J. Dean, W. Fedus, Emergent abilities of
large language models, 2022. arXiv:2206.07682.
[17] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever, Language models are unsupervised
multitask learners (2019).
[18] A. Yang, B. Yang, B. Hui, B. Zheng, B. Yu, C. Zhou, C. Li, C. Li, D. Liu, F. Huang, G. Dong, H. Wei,
H. Lin, J. Tang, J. Wang, J. Yang, J. Tu, J. Zhang, J. Ma, J. Xu, J. Zhou, J. Bai, J. He, J. Lin, K. Dang,
K. Lu, K. Chen, K. Yang, M. Li, M. Xue, N. Ni, P. Zhang, P. Wang, R. Peng, R. Men, R. Gao, R. Lin,
S. Wang, S. Bai, S. Tan, T. Zhu, T. Li, T. Liu, W. Ge, X. Deng, X. Zhou, X. Ren, X. Zhang, X. Wei,
X. Ren, Y. Fan, Y. Yao, Y. Zhang, Y. Wan, Y. Chu, Y. Liu, Z. Cui, Z. Zhang, Z. Fan, Qwen2 technical
report, arXiv preprint arXiv:2407.10671 (2024).
[19] Q. Team, Qwen2.5: A party of foundation models, 2024. URL: https://qwenlm.github.io/blog/qwen2.</p>
      <p>5/.
[20] S. Black, G. Leo, P. Wang, C. Leahy, S. Biderman, GPT-Neo: Large Scale Autoregressive
Language Modeling with Mesh-Tensorflow, 2021. URL: https://doi.org/10.5281/zenodo.5297715.
doi:10.5281/zenodo.5297715.
[21] F. Ruggeri, F. Antici, A. Galassi, K. Korre, A. Muti, A. Barrón-Cedeño, On the definition of
prescriptive annotation guidelines for language-agnostic subjectivity detection, 2023.
[22] F. Antici, F. Ruggeri, A. Galassi, K. Korre, A. Muti, A. Bardi, A. Fedotova, A. Barrón-Cedeño, A
corpus for sentence-level subjectivity detection on English news articles, in: N. Calzolari, M.-Y. Kan,
V. Hoste, A. Lenci, S. Sakti, N. Xue (Eds.), Proceedings of the 2024 Joint International Conference
on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), ELRA
and ICCL, Torino, Italia, 2024, pp. 273–285. URL: https://aclanthology.org/2024.lrec-main.25/.
[23] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, W. Chen, Lora: Low-rank adaptation
of large language models, CoRR abs/2106.09685 (2021). URL: https://arxiv.org/abs/2106.09685.
arXiv:2106.09685.
[24] H. Touvron, T. Lavril, G. Izacard, X. Martinet, M.-A. Lachaux, T. Lacroix, B. Rozière, N. Goyal,
E. Hambro, F. Azhar, A. Rodriguez, A. Joulin, E. Grave, G. Lample, Llama: Open and eficient
foundation language models, 2023. URL: https://arxiv.org/abs/2302.13971. arXiv:2302.13971.
[25] A. Q. Jiang, A. Sablayrolles, A. Mensch, C. Bamford, D. S. Chaplot, D. de las Casas, F.
Bressand, G. Lengyel, G. Lample, L. Saulnier, L. R. Lavaud, M.-A. Lachaux, P. Stock, T. L. Scao,
T. Lavril, T. Wang, T. Lacroix, W. E. Sayed, Mistral 7b, 2023. URL: https://arxiv.org/abs/2310.06825.
arXiv:2310.06825.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. V.</surname>
          </string-name>
          ,
          <article-title>The clef-2025 checkthat! lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: C.
          <string-name>
            <surname>Hauf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Kazai</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nardini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pinelli</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Silvestri</surname>
          </string-name>
          , N. Tonellotto (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2025</year>
          , pp.
          <fpage>467</fpage>
          -
          <lpage>478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Venktesh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! Lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: J.
          <string-name>
            <surname>Carrillo-de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nawrocka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ivasiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Razvan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mihail</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! lab task 1 on subjectivity in news article</article-title>
          ,
          <source>in: [4]</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Faggioli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.), Working Notes of CLEF 2025 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2025</year>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I.</given-names>
            <surname>Chaturvedi</surname>
          </string-name>
          , E. Cambria,
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Welsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Herrera</surname>
          </string-name>
          ,
          <article-title>Distinguishing between facts and opinions for sentiment analysis: Survey and challenges</article-title>
          ,
          <source>Information Fusion</source>
          <volume>44</volume>
          (
          <year>2018</year>
          )
          <fpage>65</fpage>
          -
          <lpage>77</lpage>
          . URL: https://www. sciencedirect.com/science/article/pii/S1566253517303901. doi:https://doi.org/10.1016/j. inffus.
          <year>2017</year>
          .
          <volume>12</volume>
          .006.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Casanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chanson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          , G. Faye, G. Gadek, G. Gravier,
          <string-name>
            <given-names>P.</given-names>
            <surname>Égré</surname>
          </string-name>
          , Hybrinfox at checkthat! 2024
          <article-title>- task 2: Enriching bert models with the expert system vago for subjectivity detection</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2407.03770. arXiv:
          <volume>2407</volume>
          .
          <fpage>03770</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Naveed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. U.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saqib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anwar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Usman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Barnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mian</surname>
          </string-name>
          ,
          <article-title>A comprehensive overview of large language models</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2307.06435. arXiv:
          <volume>2307</volume>
          .
          <fpage>06435</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Shokri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Filatova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Levitan</surname>
          </string-name>
          ,
          <article-title>Subjectivity detection in English news using large language models</article-title>
          , in: O.
          <string-name>
            <surname>De Clercq</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Barriere</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Barnes</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Klinger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Sedoc</surname>
          </string-name>
          , S. Tafreshi (Eds.),
          <source>Proceedings of the 14th Workshop on Computational Approaches</source>
          to Subjectivity, Sentiment, &amp;
          <article-title>Social Media Analysis, Association for Computational Linguistics</article-title>
          , Bangkok, Thailand,
          <year>2024</year>
          , pp.
          <fpage>215</fpage>
          -
          <lpage>226</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .wassa-
          <volume>1</volume>
          .17/. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .wassa-
          <volume>1</volume>
          .
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          , T. Wilson,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cardie</surname>
          </string-name>
          ,
          <article-title>Annotating expressions of opinions and emotions in language, Language Resources and Evaluation (formerly Computers and</article-title>
          the Humanities)
          <volume>39</volume>
          (
          <year>2005</year>
          )
          <fpage>164</fpage>
          -
          <lpage>210</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10579-005-7880-9.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>