<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Madrid, Spain
* Corresponding author.
$ selina.meyer@utn.de (S. Meyer); michael.roth@utn.de (M. Roth)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>nlu@utn at CheckThat! 2025: Combining Bias Sensitivity, Linguistic Features, and Persuasion Cues in an Ensemble for Subjectivity Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Selina Meyer</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Roth</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Technology Nuremberg</institution>
          ,
          <addr-line>Dr.-Luise-Herzberg-Straße 4, 90461 Nürnberg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This paper details our participation in task 1 on subjectivity detection at 2025's CheckThat! lab. Specifically, we focused on monolingual English data provided by the task organizers. Our approach consisted of an ensemble model, made up of a) a classifier pretrained on political bias data and fine-tuned on the subjectivity training data, b) a BERT classifier fine-tuned on linguistically augmented training data, and c) a BERT classifier fine-tuned on training data augmented with persuasion techniques used in the text. The final label was determined using majority voting. Our approach resulted in a Macro F1 score of 74.86% on the test set, ranking 7ℎ place on the task leaderboard. Based on additional experiments conducted over the course of our participation, our ensemble-based system outperformed state-of-the-art large language models, including Google's Gemini-Flash and GPT-4o-Mini on this task, highlighting that, given the presented experimental setup, this type of task is still a challenge for generative AI.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;subjectivity detection</kwd>
        <kwd>credibility</kwd>
        <kwd>natural language processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Subjective writing in news can be a marker of partisan and biased reporting but can also be a stylistic
and narrative device to increase the readability of articles [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. udging subjectivity is no easy task—even
for human experts—as it is shaped by subjective interpretation. Nevertheless, Antici et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] have
introduced a corpus of sentences from news articles judged as objective and subjective, garnering high
inter-annotator agreement through a multistep approach. Their corpus consists of articles sourced
from British news outlets with diverse political orientations and annotators were tasked to annotate
on a sentence-to-sentence basis, without being given context for the respective sentences. In our
participation in the English subtask of Task 1 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] presented at CheckThat! 2025 [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], we draw on the
insights from related research on news credibility to improve classifier performance.
      </p>
      <p>
        According to Antici et al.’s definition used for annotation, a “sentence is considered subjective when
it is based on—or influenced by—personal feelings, tastes, or opinions” [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This betrays a close relation
to the concept of credibility, which is often defined by research-based credibility cues in the context
of natural language processing [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Such cues can be context-based, but also content-based, meaning
identifiable on the basis of the text itself. Content-based credibility signals include partisanship and
bias, linguistic cues, general quality characteristics of text and clickbaitiness of headlines, presence of
logical fallacies, and use of persuasion techniques [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Earlier work, especially before the introduction of
transformer-based models, has also heavily focused on the use of lexical and other linguistic information
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Other submissions to this task in previous years have used linguistic features, and most have used
transformer models as underlying architectures [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Casanova et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], who ranked first place in 2024’s
edition of this task, used an ensemble approach based on two transformer models, as well as subjectivity
scores based on an independently developed expert system for subjectivity and vagueness detection.
      </p>
      <p>
        In our approach to the task, we build on the previously seen success of using ensemble models for
subjectivity detection. In contrast to Casanova et al.’s [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] approach, each of the three models in our
ensemble is informed by a diferent aspect of cues which are likely to inform subjectivity detection
classifiers. Specifically, we leverage a model trained on the related task of persuasion technique detection
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] to augment the training data before fine-tuning a BERT-instance, fine-tune a BERT-based classifier
previously fine-tuned on a bias detection task [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], and combine this with a third BERT-instance
ifne-tuned on augmented data based on our own linguistic analysis of the provided training data. Our
ensemble model reaches a Macro F1 score of 74.86% on the test set, ranking 7ℎ place on the oficial task
leaderboard, with only 0.58% separating us from 5ℎ place. This is an improvement of 21 percentage
points over the baseline provided by the task organizers.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methods</title>
      <p>Below, we describe the two data augmentation approaches we employed for this task, as well as the
classifiers and fine-tuning involved in creating the ensemble model.</p>
      <sec id="sec-2-1">
        <title>2.1. Linguistic Analysis and Augmentation</title>
        <p>
          Gajewska [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] explored the role of pronouns and modal verbs, as well as emotional and polarising words
in subjectivity detection, comparing the predictive power of naive models trained with these features
with BERT-based methods. In their experiments, simple BERT-finetuning on the training data, with no
additional meta-features, outperformed all other approaches tested in their experiments, resulting in a
Macro F1 score of 0.70 on the test-set. This indicates, that a naive approach to linguistic feature selection
is not enough to successfully inform subjectivity classification. Instead, we analyze the training data to
identify linguistic features which statistically diferentiate subjective from objective sentences.
        </p>
        <p>
          To achieve this, we tokenize the training data using spaCy and parse the simple and detailed
part-ofspeech (POS) tags, syntactic dependencies, and named entity labels for each token [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Following this,
we identify significant diferences in the frequency of each of these features between the subjective
and objective class. Table 1 shows all features with significantly difering frequency between the two
classes, and their corresponding bonferroni-corrected p-values.
        </p>
        <p>Based on these insights, the original sentences in the training data were augmented with a [SEP]
token after the original sentence, followed by the numerical count of each diferentiating feature in the
sentence (see Table 2).</p>
        <sec id="sec-2-1-1">
          <title>1Code available at https://github.com/SelinaMeyer/nlu-utn_at_CheckThat2025</title>
          <p>Linguistically-informed augmentation
Gone are the days when they led the world in recession-busting[SEP]ADJ-count: 0; ADV-count: 0; AUX-count:
1; NUM-count: 0; PRON-count: 1; PROPN-count: 0; CD-count: 0; JJ-count: 0; MD-count: 0; NNP-count: 0;
PRP-count: 1; RB-count: 0; VB-count: 0; VBD-count: 1; advmod-count: 1; compound-count: 1; nummod-count: 0;
quantmod-count: 0; DATE-count: 0; ORG-count: 0; PERCENT-count: 0; ’
Persuasion technique-based augmentation
Gone are the days when they led the world in recession-busting[SEP][’Loaded Language’][SEP][[0,
62]][SEP][’0.83782977’]</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Persuasion Technique-Based Augmentation</title>
        <p>
          Subjectivity in writing can manifest in the use of loaded language, name calling, appeal to values, use
of faulty argumentation, e.g. strawman or red herring arguments, and other types of biased writing.
These techniques have been explored by other researchers in the context of persuasion and propaganda
detection [
          <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
          ]. To explore whether knowledge of such techniques can enhance the performance of
subjectivity classifiers, we used a model created in the context of a shared task in SemEval 2023 [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]
available via API access on GATE Cloud2. The model is passed a piece of text and returns the persuasion
techniques detected in the sentence, along with confidence scores and the start and end character of
the span for which a respective persusasion technique was identified. We augment the training data
with the returned information by appending it with a list of identified persuasion techniques, followed
by the list of spans and confidence scores, each separated by [SEP] tokens (See Table 2).
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Model Training</title>
        <p>
          Our ensemble method employs three models: BERTpers is fine-tuned on the persuasion-augmented
training data, BERTling is fine-tuned on the linguistically-augmented training data. Both of these are
based on instances of bert-base-uncased [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]3. In contrast, BERTbias is based on a pre-trained
BERT model available on Hugging Face, as provided by Baly et al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]4, which was initially fine-tuned
on news articles labeled with their displayed political ideology (right, center, left). For the subjectivity
detection task, we further fine-tuned this model using the non-augmented training data. Since the
original classifier was designed for a multi-class task with three categories, we adapted the classification
head to accommodate binary classification. After this modification, we froze the weights of the initial
layers and only fine-tuned the final four layers. In the resulting ensemble model, the class of a text
sample is decided via simple majority voting between the three models.
        </p>
        <p>We use the same hyperparameter configuration for Bertling and Bertpers during fine-tuning, applying
a learning rate of 2e-5, weight decay of 0.01, a batch size of 16 and training over 20 epochs. To fine-tune
bertbias we reduce the learning rate to 1e-5. An early stopping callback was implemented for all models
with a patience of three epochs. The best model, based on performance on the dev-split was loaded at
the end of training. Fine-tuning was performed on a NVIDIA A40 GPU and took between 74.68 and
125.83 seconds per model. The early stopping callback resulted in a training of 11 epochs for BERTbias,
10 epochs for BERTpers, and 6 epochs for BERTling.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>2https://cloud.gate.ac.uk/shopfront/displayItem/persuasion-classifier 3https://huggingface.co/google-bert/bert-base-uncased 4https://huggingface.co/bucketresearch/politicalBiasBERT</title>
        <p>The ensemble approach led to a Macro F1 increase of 7% over the baseline on the dev split, 8% on
the dev-test split, and 21% on the test split as returned by the scorer provided by the task organizers.
Performance on the test set as displayed on the oficial leaderboard was at Macro F1 of 74.86% placing
our submission in 7ℎ place.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Ablation Studies</title>
      <p>We conduct a series of ablation studies to assess the necessity of the individual classifiers in the
ensemble, compare the performance of our ensemble model with state-of-the-art large language models
(LLMs)—namely Google Gemini, Google Gemma, and GPT-4—and explore an LLM-in-the-loop approach
to confirm or reject the predictions made by the ensemble.</p>
      <sec id="sec-4-1">
        <title>4.1. Replacing Individual Classifiers</title>
        <p>To test the role the individual classifiers in the ensemble play in classification performance, we replace
each classifier in turn with a bert-base-uncased instance, fine-tuned on the non-augmented training
data (BERTbaseline). The results are presented in Table 4.</p>
        <p>Interestingly, although BERTbias alone leads to the lowest classification results of all the classifiers
included in our ensemble model, replacing it with BERTbaseline yields the largest discrepancy in
classification results compared to the original ensemble on the dev split, whereas removing BERTpers, which
on its own yields the highest score on the dev split, does not decrease the classification score on the dev
split at all compared to the ensemble results and leads to the lowest decrease in Macro F1 on the test split.
On the dev-test split, replacing BERTbias with BERTbaseline leads to a Macro F1 increase of 1%, whereas
replacing BERTling decreases the score by the same amount and on the test split, replacing any of the
three classifiers in the ensemble leads to a decrease in classification scores, with the diference being
smallest when Bertpers is replaced. To explore potential reasons for this, we calculate the agreement
rate between the individual classifiers on the test set, presented in Table 5.</p>
        <p>The classifications produced by BERTpers largely align with those of BERTbaseline, which accounts
for the minimal change in results when the two models are substituted for one another. At the same
time, BERTling and BERTbias show larger discrepancies with the remaining models, suggesting they
contribute more variability to the predictions. This added diversity may explain why replacing them
with BERTbaseline results in larger discrepancies in classification scores for the dev and test splits.</p>
        <p>BERTbaseline</p>
        <p>BERTbias
BERTpers
BERTling</p>
        <p>BERTbias
80%
1</p>
        <p>BERTpers
91.33%
78%
1</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Generative LLM-based Classification</title>
        <p>
          Given that the data at hand is hard to annotate even for humans [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], we were interested in
how generative LLMs perform at this task. We conducted a range of experiments employing
two proprietary LLMs, namely Google’s gemini-2.0-flash-lite (Gemini Flash) and OpenAI’s
gpt-4o-mini-2024-07-18 (GPT-4o-mini), as well as the smaller, open-source LLM gemma2-9b-it
(Gemma2-9b), also created by Google, which can be run on a single GPU. We test diferent prompt
conditions, inspired by the annotation guidelines and examples used for the training data provided by
Antici et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. To save resources, the full range of experiments was conducted only with Gemini-Flash
on the dev split. The prompt yielding the best classification results utilizing Gemini-Flash was then
also run using Gemma2-9b to explore whether smaller, locally run models can compare to large scale
model performance on this task, and GPT-4o-mini to identify whether using a diferent model with
comparable size can boost performance. Finally, the best performing system was used to predict on the
dev-test and test splits.
        </p>
        <p>The prompts provided for the models were each made up of a system prompt, telling the model that it
is an expert at judging the subjectivity of texts, followed by diferent prompt variants explained below,
and finally an instruction to provide the response in JSON format, containing a label, an explanation for
why the label was given, the text fragment that was most important for the decision, and, depending on
the prompt variant, the subjectivity or objectivity subclass.</p>
        <p>
          The following prompt variants were explored:
• Simple prompt - The model was instructed to classify sentences as subjective or objective, with no
detailed definitions of what constitutes subjectivity or objectivity, to identify the model’s inherent
ability to diferentiate between subjective and objective texts.
• Full guidelines - The full guidelines provided in Antici et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], including the definitions of
diferent subclasses of subjectivity and objectivity, were adapted to the use of LLMs and passed
to the model.
• Few-shot-style - The examples for various objectivity and subjectivity subclasses provided in the
guidelines by Antici et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] were restructured into a few-shot learning format and passed to the
model.
• Guidelines without examples - All examples provided in the guidelines were removed, and only
the definitions of each subjective and objective subclass were passed to the model.
        </p>
        <p>The exact wording for each of the conditions is provided in Appendix A. The results of this range of
experiments are presented in Table 6.</p>
        <p>In the Full guidelines condition, Gemini-Flash achieved classification performance on the dev split,
which had been used as a validation set during fine-tuning of the BERT-based models, that closely
approached that of the ensemble model. This initially suggested that the LLM-based system might
outperform the ensemble classifier on the dev-test and test splits. However, this expectation was not
met. On the dev-test split, Gemini-Flash’s Macro F1 score was 6 percentage points lower than that of the
ensemble, suggesting that the dev-test split is generally more challenging to classify than the remaining
splits. This assumption is further supported by the consistently higher results yielded by tested models
on the test split relative to the dev-test split. On the test split, Gemini-Flash’s classification score was 3
percentage points lower than the ensemble’s result.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Generative LLM-in-the-Loop</title>
        <p>Finally, we test whether generative LLMs could be used to improve ensemble-based classifications. To
this end, we pass Gemini-Flash the annotation guidelines (as in the full guidelines condition above)
along with a sample classified by the ensemble model and information about the model’s performance
and the label distribution found in the data. The model is then prompted to confirm or reject the
ensemble model’s classification, based on the annotation guidelines. Since an error analysis of the
LLM experiments above revealed that a frequent error source was the model’s inability to diferentiate
between the author’s opinion and that of a quoted third-party, a sentence was added to the instructions
to increase awareness of this. This approach led to a decrease in classification scores compared to using
only the ensemble model across the dev, dev test and test splits, leading to Macro F1 scores of 0.75, 0.68,
and 0.72, respectively.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Error Analysis and Discussion</title>
      <p>As mentioned above, our error analysis of generative LLM classifications showed that Gemini-Flash
struggled to diferentiate between the opinions of the author and those of third parties quoted in the text
when prompted using the annotation guidelines. An analysis of misclassified samples of the ensemble
model in the test split revealed a bias towards the subjective class. 68.18% of misclassifications were
false positives.</p>
      <p>Based on a closer analysis of the misclassified samples in the test set, it appears that many of the
sentences constitute observations about a prominent person or event, as well as personal narratives
either based on the author’s experiences or not discernible as quotations in the sentence itself (see
examples in Table 7). Although not directly subjective, these samples appear to exhibit a narrative style,
which could make them harder to classify.</p>
      <p>
        Although our approach did not achieve top performance compared to other submissions, we draw
a couple of interesting conclusions from our experiments and ablation studies, which we discuss
here. The first is that, although generative LLMs have led to many advancements in natural language
understanding, subjectivity detection given the current experimental setup still poses a challenge even
to top-performing LLMs, such as GPT-4 and Google Gemini. The task requires a lot of nuance and
is hard to annotate even for humans [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which appears to have limited LLM performance in our
experiments. We also find indications that Gemini-Flash’s internal representation of subjectivity may
difer significantly from the guidelines developed by Antici et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This is reflected in the remarkably
low performance of the simple prompt condition in our LLM experiments, where the model judged
subjectivity based solely on its own understanding. Future work could investigate this discrepancy
more systematically, for example by explicitly probing the model’s internal criteria for subjectivity
across diverse contexts and comparing diferent LLMs.
      </p>
      <p>
        Furthermore, our ensemble method achieved better classification scores independently than in
combination with the generative LLM-in-the-loop. This suggests that for certain classification tasks
and experimental setups, in which generative LLMs do not perform particularly well on their own,
incorporating them in this manner does not lead to improvements. A possible direction for future work
is testing out the capabilities of LLMs in this context more extensively, for instance through dynamic
selection of few-shot samples, as presented for example in An et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] or other state-of-the-art
prompting approaches, which might significantly increase LLM-performance for this task.
      </p>
      <p>
        We also find that explicitly providing linguistic information relevant to a specific dataset, as we did
in the training of the BERTling classifier, can be beneficial for classification performance, even though
transformer-based models already implicitly encode such information [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], underlined by the results of
the replacement ablation study, which showed that BERTling appeared to be the most impactful classifier
in the ensemble. The positive impact of leveraging linguistic features, as operationalized by POS-tags
and similar automatically retrievable information, for judging subjectivity has been shown by Nakashole
and Mitchell [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] in the past, although not in combination with transformer-based approaches. Future
work could explore whether supplying this type of information to generative LLMs might improve
classification results.
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this paper, we presented our approach to the English subtask of Task 1 of CheckThat!2025. Our
ensemble approach used three BERT-based models, each informed by linguistic diferences between the
subjective and objective classes in the training data, research on persuasion techniques in propaganda,
and bias in news articles. The system placed in the upper third of the English leaderboard.</p>
      <p>Our ablation studies showed that using linguistic information in an explicit way contributed to the
ensemble and that models, which perform poorly by themselves can still contribute to an ensemble
approach, as was the case for BERTBIAS in our experiments. Experiments with generative LLM-based
classification and LLM-in-the-loop methods showed that the task remains a challenge for generative AI,
given the presented experimental setup.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used GPT-4 in order to: Paraphrase and reword,
improve writing style. After using this tool, the authors reviewed and edited the content as needed and
take full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-8">
      <title>A. Prompts</title>
      <sec id="sec-8-1">
        <title>A.1. System Prompt</title>
        <p>You are an expert at judging the subjectivity of texts and are skilled at
adhering to guidelines provided to you.</p>
        <p>For the text provided to you, first judge whether it is objective or subjective.
Then compare your initial judgement to each of the provided objectivity and
subjectivity criteria to determine whether your initial judgement is correct.
If your initial judgement is correct, return the label you have chosen.
If your initial judgement is incorrect, return the label that is correct
according to the guidelines.</p>
      </sec>
      <sec id="sec-8-2">
        <title>A.2. Simple Prompt</title>
        <p>You are tasked with judging whether a sentence is subjective or objective.
Give your answer in a JSON dictionary with the following format:
"explanation": [a short explanation of why the sentence is subjective or
objective]
"decisive text fragment": [the text snippet that is the most important for
deciding between objective and subjective]
"label": [SUBJ for subjective or OBJ for objective]
Judge the following sentence:
{text}</p>
      </sec>
      <sec id="sec-8-3">
        <title>A.3. Full Guidelines</title>
        <p>You are tasked with judging whether a short text is subjective or objective.
You are handed the guidelines below, given in triple backticks, to inform
your decision.</p>
        <p>Guidelines:
```
{guidelines (Appendix A.7)}
```
Give your response in the following JSON-Format:
"label": [SUBJ for subjective or OBJ for objective],
"explanation": [a short explanation, why the sentence is objective or
subjective],
"decisive text fragment": [the text snippet that is the most important for
your decision],
"case": [The specific case of objectivity or subjectivity],
Judge the following sentence according to the guidelines provided above:
{text}</p>
      </sec>
      <sec id="sec-8-4">
        <title>A.4. Guidelines without Examples</title>
        <p>You are tasked with judging whether a short text is subjective or objective.
You are handed the guidelines below, given in triple backticks, to inform
your decision.</p>
        <p>Guidelines:
```
Definitions
**Subjective.** A sentence is considered subjective when it is based on
—or influenced by— personal feelings, tastes, or opinions. Otherwise, the
sentence is considered objective.
**Specific Subjective Cases**
SUBJ 1.</p>
        <p>A sentence is subjective if it explicitly reports the personal opinion of its
author. Rhetorical questions are considered as an expression of an opinion
as well; see Ex. (c). Additionally, speculations which draw conclusions are
considered as opinions.</p>
        <p>SUBJ 2.</p>
        <p>A sentence is subjective if it contains sarcastic or ironic expressions
attributable to its author.
.... (see Appendix A.7 for remaining categories)
```
Give your response in the following JSON-Format:
"label": [SUBJ for subjective or OBJ for objective]
"explanation": [a short explanation, why the sentence is objective or
subjective],
"decisive text fragment": [the text snippet that is the most important for
your decision],
"case": [The specific case of objectivity or subjectivity],
Judge the following sentence according to the guidelines provided above:</p>
      </sec>
      <sec id="sec-8-5">
        <title>A.5. Few-Shot Style</title>
        <p>You are tasked with judging whether a sentence is subjective or objective.
Guidelines:
```
Definitions
**Subjective.** A sentence is considered subjective when it is based on
—or influenced by— personal feelings, tastes, or opinions. Otherwise, the
sentence is considered objective.
**Objective.** If a sentence does not meet any subjectivity type listed in
the previous section, it is considered objective. Here we include examples
of objective sentences which may be wrongly interpreted as subjective.
Text: India, who was the bridesmaid at the King’s wedding to Princess
Diana in 1981, could not be seen in the footage, but filmed the video as she
walked through the grounds of the royal residence.</p>
        <p>Label: Obj
Explanation: The sentence is objective because the author describes a
historical event without giving any opinion or personal comment
Text: It is a sad truth that many of the villages in this region of
Portugal are dying.</p>
        <p>Label: Subj
Explanation: The sentence is subjective because the author explicitly conveys
their personal emotions, making the sentence subjective.
...(see Appendix A.7 for remaining categories)
```
Answer in the following JSON-Format:
"explanation": [a short explanation of why the sentence is subjective or
objective],
"decisive text fragment": [the text snippet that is the most important for
deciding between objective and subjective],
"label": [SUBJ for subjective or OBJ for objective],
Judge the following sentence:
{text}</p>
      </sec>
      <sec id="sec-8-6">
        <title>A.6. LLM-as-a-Judge</title>
        <p>You are tasked with judging whether the subjectivity prediction of a short
text given by a classifier with 80% accuracy is correct or not.
The data contains around twice as many objective sentences as subjective
sentences.</p>
        <p>You are handed the guidelines below, given in triple backticks, to inform
your decision.</p>
        <p>Guidelines:
Be mindful of the presence of quotation marks and other markers that
indicate the presence of a third-party opinion.</p>
        <p>Such markers should be considered as explicit third-party opinions and make
the sentence objective, even if they are not clearly stated in the sentence.
The texts given to you for judgement are labelled as SUBJ for subjective and
OBJ for objective.</p>
        <p>Given the guidelines above, judge the correctness of the subjectivity label
of the following Text:
Text: {text}
Label: [label]
Give your response in the following JSON-Format:
"explanation": [a short explanation of why the label is correct or incorrect,
referring to the guidelines],
"correct": [True if the label is correct, False if it is incorrect]</p>
      </sec>
      <sec id="sec-8-7">
        <title>A.7. Guidelines, based on [2]</title>
        <p>Definitions
**Subjective.** A sentence is considered subjective when it is based on
—or influenced by— personal feelings, tastes, or opinions. Otherwise, the
sentence is considered objective.</p>
        <p>The following Sentence is objective because the author describes a historical
event without giving any opinion or personal comment:
Ïndia, who was the bridesmaid at the King’s wedding to Princess Diana in 1981,
could not be seen in the footage, but filmed the video as she walked through
the grounds of the royal residence.¨
In contrast, in the next Sentence the author explicitly conveys their personal
emotions, making the sentence subjective:
Ït is a sad truth that many of the villages in this region of Portugal are
dying.¨
**Specific Subjective Cases**
SUBJ 1.</p>
        <p>A sentence is subjective if it explicitly reports the personal opinion of its
author. Rhetorical questions are considered as an expression of an opinion
as well; see Ex. (c). Additionally, speculations which draw conclusions are
considered as opinions, see Ex. (d).</p>
        <p>Examples:
(a) It has everything you could want in a holiday: beautiful sandy beaches
and clear waters, ancient history and culture, delicious food (the Greek
salads are simply on another level), island hopping, nightlife and more.
(b) After treading vineyard soils and seeing grapes ripening, that merlot
becomes more than just a Wednesday night relaxant.
(c) Do they really think other nations sprouted up out of the ground?
(d) But Putin will hope to sow uncertainty in the eyes of policymakers’
meetings in New York.</p>
        <p>SUBJ 2.</p>
        <p>A sentence is subjective if it contains sarcastic or ironic expressions
attributable to its author.</p>
        <p>Examples:
(e) It’s no lie that the USA is one heck of a big country (said in a southern
twang).
(f) With Land Rover bowdlerising images of the vehicle into little more than
a perfume advertisement on TV[...].
(g) Especially if you’re more excited at the prospect of sampling rare bottles
from the cellar than snapping vineyard selfies.</p>
        <p>SUBJ 3.</p>
        <p>A sentence is subjective if it contains exhortations of personal auspices
made by its author.</p>
        <p>Examples:
(h) The West should arm Ukraine faster.</p>
        <p>SUBJ 4.</p>
        <p>A sentence is subjective if it contains discriminating or downgrading
expressions.</p>
        <p>Examples:
(i) And what is even more evident is the perverse role reversal that is
taking place, in which he who sits in Rome has the task of formulating
heterodox principles opposed to Catholic doctrine, and his accomplices in the
Dioceses have the role of scandalously applying them, in an infernal attempt
to undermine the Moral law in order to obey the spirit of the world.
(j) How did we reach the stage where priests and bishops cowered like
frightened puppies before a common flu, where their predecessors ministered
fearlessly among the lepers, the cripples, and the victims of typhoid,
cholera, smallpox, and Bubonic Plague?
SUBJ 5.</p>
        <p>A sentence is subjective if it contains rhetorical figures, like hyperboles,
explicitly made by its author to convey their opinion.</p>
        <p>Examples:
(k) Barcelona where it all began, Messi was a king in Catalonia and he lived
like one too.
(l) The churches, and the Catholic Church in particular (which is by far
the largest), had the ability to put an end to the lockdown madness and the
COVIDterror campaign, had they wished to do so.
(m) So it must be biochemistry that is really what is racist.
**Specific Cases of Objectivity**
If a sentence does not meet any subjectivity type listed in the previous
section, it is considered objective. Here we include examples of objective
sentences which may be wrongly interpreted as subjective.</p>
        <p>Case 1.</p>
        <p>A sentence is objective when it reports on news or historical facts that are
quoted by the author of the sentence.</p>
        <p>Examples:
(a) President Putin has just reiterated his threat to use nuclear weapons and
announced that Russiancontrolled Ukrainian territory will become part of the
Russian Federation. (b) In the modern era electroconvulsive therapy, first
used in 1938, became a treatment for some serious forms of depression in the
post-war decades.</p>
        <p>Case 2.</p>
        <p>A sentence is objective when it describes the personal feelings, emotions or
moods of the writer, without conveying opinions on other matters.
Examples:
(c) I was definitely surprised at how emotional I felt watching the service.
(d) The second I saw him, I felt a jolt of connection.</p>
        <p>Case 3.</p>
        <p>A sentence is objective if it expresses an opinion, claim, emotion or a point
of view that is explicitly attributable to a third-party (e.g., a person
mentioned in the text).</p>
        <p>Examples:
(e) Frank Drake believed that the universe had to contain other intelligent
beings.
(f) Ÿou showed callous indifference to Dean’s fate after he had been repeatedly
stabbedt¨he judge said.</p>
        <p>Note: The presence of quotation marks (“ ”), when used to quote a third
person (be it at the beginning
of the sentence, at the end, or both), represents an explicit third-party
opinion, even if it is not clearly stated in the sentence.</p>
        <p>Examples:
(g) “Crosbie is an extremely violent man who has no place in society, and we
welcome the jury’s verdict today.”
(h) “My children have lost their hero and I have lost my chosen person - the
person I chose to spend my life with.
(i) For these reasons and out of conviction, I consider myself bound in my
conscience to say no.”
Case 4.</p>
        <p>A sentence is objective if it contains a comment made by the author of the
sentence that does not draw any conclusion.</p>
        <p>In particular, the author doesn’t convey their personal interpretation or
opinion, leaving the discussion on the topics of interest open.
Examples:
(j) It is not clear yet which of the couples from the E4 reality show remain
together and who have now, because the series has not concluded.
(k) Do car manufacturers know how far their EVs will really go?
(l) Exact figures are hard to come by, but Ukraine may well have more troops
available than Russia now.</p>
        <p>Case 5.</p>
        <p>A sentence is objective if it contains factual conclusions made by the author
of the sentence that do not convey any stance or personal opinion, or are
justified up by a non-personal hypothesis.</p>
        <p>Examples:
(m) In years gone by, travel to Japan was notoriously expensive, but the
devaluing of the yen has made it more accessible.
(n) The bottom-up approaches which target the molecular, genetic and
electrical fundamentals of the brain can assist top-down approaches to brain
disorder such as talking therapies.
(o) Based on our experiences and road tests, a good rule of thumb is to expect
to achieve somewhere between 75 and 80 per cent of a car’s WLTP Combined
range[. . . ]</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Steensen</surname>
          </string-name>
          ,
          <article-title>Subjectivity as a journalistic ideal</article-title>
          , in: B.
          <string-name>
            <surname>K. Fonn</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Hornmoen</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Hyde-Clarke</surname>
            ,
            <given-names>Y. B.</given-names>
          </string-name>
          <string-name>
            <surname>Hågvar</surname>
          </string-name>
          (Eds.),
          <article-title>Putting a Face on it: Individual Exposure and</article-title>
          Subjectivity in Journalism, Cappelen Damm Akademisk,
          <year>2017</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>47</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Antici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bardi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fedotova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <article-title>A corpus for sentence-level subjectivity detection on English news articles</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            , M.-
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Kan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sakti</surname>
          </string-name>
          , N. Xue (Eds.),
          <source>Proceedings of the 2024 Joint International Conference on Computational Linguistics</source>
          ,
          <article-title>Language Resources and Evaluation (LREC-COLING 2024), ELRA</article-title>
          and
          <string-name>
            <given-names>ICCL</given-names>
            ,
            <surname>Torino</surname>
          </string-name>
          , Italia,
          <year>2024</year>
          , pp.
          <fpage>273</fpage>
          -
          <lpage>285</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .lrec-main.
          <volume>25</volume>
          /.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaghouani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nawrocka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ivasiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Razvan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mihail</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! lab task 1 on subjectivity in news article</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.), Working Notes of CLEF 2025 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2025</year>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. V.</surname>
          </string-name>
          ,
          <article-title>The clef-2025 checkthat! lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: C.
          <string-name>
            <surname>Hauf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Kazai</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nardini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pinelli</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Silvestri</surname>
          </string-name>
          , N. Tonellotto (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2025</year>
          , pp.
          <fpage>467</fpage>
          -
          <lpage>478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hafid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schellhammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Setty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundriyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Todorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Venktesh</surname>
          </string-name>
          ,
          <article-title>Overview of the CLEF-2025 CheckThat! Lab: Subjectivity, fact-checking, claim normalization, and retrieval</article-title>
          , in: J.
          <string-name>
            <surname>Carrillo-de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>I.</given-names>
            <surname>Srba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Razuvayevskaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Leite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. B.</given-names>
            <surname>Schlicht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tonelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>García</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Lottmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Teyssou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Porcellini</surname>
          </string-name>
          , et al.,
          <article-title>A survey on automatic credibility assessment of textual credibility signals in the era of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2410.21360</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          , G. Da San Martino,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the clef-2024 checkthat! lab: check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>28</fpage>
          -
          <lpage>52</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Casanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chanson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          , G. Faye, G. Gadek, G. Gravier,
          <string-name>
            <given-names>P.</given-names>
            <surname>Égré</surname>
          </string-name>
          , Hybrinfox at checkthat! 2024
          <article-title>-task 2: Enriching bert models with the expert system vago for subjectivity detection</article-title>
          ,
          <source>in: CLEF 2024-Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>B.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Razuvayevskaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Heppell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Leite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scarton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          , ShefieldVeraAI at SemEval
          <article-title>-2023 task 3: Mono and multilingual approaches for news genre, topic and persuasion technique classification</article-title>
          , in: A.
          <string-name>
            <surname>K. Ojha</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Doğruöz</surname>
            , G. Da San Martino, H. Tayyar Madabushi,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , E. Sartori (Eds.),
          <source>Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>1995</fpage>
          -
          <lpage>2008</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .275/. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .
          <fpage>275</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Baly</surname>
          </string-name>
          , G. Da San Martino, J. Glass,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>We can detect your bias: Predicting the political ideology of news articles</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <source>EMNLP '20</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>4982</fpage>
          -
          <lpage>4991</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <article-title>Political bias classification using finetuned bert model (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>E.</given-names>
            <surname>Gajewska</surname>
          </string-name>
          , Eevvgg at checkthat!
          <year>2024</year>
          <article-title>: evaluative terms, pronouns and modal verbs as markers of subjectivity in text</article-title>
          , Faggioli et al.[
          <volume>22</volume>
          ] (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Honnibal</surname>
          </string-name>
          , I. Montani,
          <string-name>
            <given-names>S. Van</given-names>
            <surname>Landeghem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Boyd</surname>
          </string-name>
          , spaCy: Industrial-strength
          <source>Natural Language Processing in Python (</source>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>G.</given-names>
            <surname>Da San Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Petrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <article-title>Fine-grained analysis of propaganda in news articles</article-title>
          , in: K. Inui,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          Wan (Eds.),
          <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>5636</fpage>
          -
          <lpage>5646</lpage>
          . URL: https://aclanthology.org/D19-1565/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1565.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Piskorski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stefanovitch</surname>
          </string-name>
          , G. Da San Martino, P. Nakov, SemEval
          <article-title>-2023 task 3: Detecting the category, the framing, and the persuasion techniques in online news in a multi-lingual setup</article-title>
          , in: A.
          <string-name>
            <surname>K. Ojha</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Doğruöz</surname>
            , G. Da San Martino, H. Tayyar Madabushi,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          , E. Sartori (Eds.),
          <source>Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>2343</fpage>
          -
          <lpage>2361</lpage>
          . URL: https: //aclanthology.org/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .317/. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .semeval-
          <volume>1</volume>
          .
          <fpage>317</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/N19-1423/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          , J.-G. Lou,
          <article-title>Skill-based few-shot selection for in-context learning</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>13472</fpage>
          -
          <lpage>13492</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .emnlp-main.
          <volume>831</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>831</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jawahar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sagot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Seddah</surname>
          </string-name>
          ,
          <article-title>What does BERT learn about the structure of language?</article-title>
          , in: A.
          <string-name>
            <surname>Korhonen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Traum</surname>
          </string-name>
          , L. Màrquez (Eds.),
          <article-title>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>3651</fpage>
          -
          <lpage>3657</lpage>
          . URL: https://aclanthology.org/P19-1356/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P19</fpage>
          -1356.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N.</given-names>
            <surname>Nakashole</surname>
          </string-name>
          , T. Mitchell,
          <article-title>Language-aware truth assessment of fact candidates, in: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2014</year>
          , pp.
          <fpage>1009</fpage>
          -
          <lpage>1019</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>