<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the CLEF-2025 CheckThat! Lab Task 1 on Subjectivity in News Articles</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Federico Ruggeri</string-name>
          <email>federico.ruggeri6@unibo.it</email>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arianna Muti</string-name>
          <email>arianna.muti@unibocconi.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Katerina Korre</string-name>
          <email>k.korre@athenarc.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julia Maria Struß</string-name>
          <email>julia.struss@f-potsdam.de</email>
          <xref ref-type="aff" rid="aff7">7</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Melanie Siegel</string-name>
          <email>melanie.siegel@h-da.de</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Wiegand</string-name>
          <email>michael.wiegand@univie.ac.at</email>
          <xref ref-type="aff" rid="aff9">9</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Firoj Alam</string-name>
          <email>rojalam@gmail.com</email>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Md. Raul Biswas</string-name>
          <email>mbiswas@hbku.edu.qa</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wajdi Zaghouani</string-name>
          <email>wajdi.zaghouani@northwestern.edu</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Nawrocka</string-name>
          <xref ref-type="aff" rid="aff10">10</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bogdan Ivasiuk</string-name>
          <email>bogdan.ivasiuk@studio.unibo.it</email>
          <xref ref-type="aff" rid="aff8">8</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gogu Razvan</string-name>
          <email>gogu.razvan2001@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreiana Mihail</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Athena RC</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Bocconi University</institution>
          ,
          <addr-line>Milan</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Bucharest University</institution>
          ,
          <country country="RO">Romania</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Darmstadt University of Applied Sciences</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Hamad Bin Khalifa University</institution>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Northwestern University in Qatar</institution>
          ,
          <addr-line>Doha</addr-line>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>Qatar Computing Research Institute</institution>
          ,
          <addr-line>Doha</addr-line>
          ,
          <country country="QA">Qatar</country>
        </aff>
        <aff id="aff7">
          <label>7</label>
          <institution>University of Applied Sciences Postdam</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff8">
          <label>8</label>
          <institution>University of Bologna</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff9">
          <label>9</label>
          <institution>University of Vienna</institution>
          ,
          <country country="AT">Austria</country>
        </aff>
        <aff id="aff10">
          <label>10</label>
          <institution>Warsaw University</institution>
          ,
          <country country="PL">Poland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>We present an overview of Task 1 of the eighth edition of the CheckThat! lab at the 2025 edition of the Conference and Labs of the Evaluation Forum (CLEF). The task required participants to determine whether individual sentences from news articles expressed subjective viewpoints, such as opinions or personal bias, or presented objective, fact-based information. The task was o‌ered in nine languages: Arabic, Bulgarian, English, German, Italian, Greek, Polish, Romanian, and Ukrainian, as well as in a multilingual setting. We curated datasets for each language, comprising roughly 14,000 sentences sourced from diverse news outlets. Participants were tasked with developing classication systems to identify subjectivity (personal opinions or biases) and objectivity (factual information) at the sentence level. A total of 22 teams participated in the task, submitting 436 valid runs across all language tracks. Most systems were based on transformer models, with approaches ranging from ne-tuning language-specic and multilingual encoders to applying English-centric models in combination with machine translation. Several teams also experimented with ensemble techniques, handcrafied features, and in-context learning using large language models. Systems were evaluated using macro-averaged F1 score to ensure equal weighting of subjective and objective classes. Performance varied considerably by language: German, Italian, English and Romanian yielded the highest results. In contrast, Greek and Ukrainian emerged as the most challenging languages, with no team surpassing the 0.65 and 0.51 F1 score marks, respectively. Task 1 o‌ers a valuable benchmark for the development and evaluation of multilingual subjectivity detection systems. This paper presents an overview of Task 1, including datasets, system strategies, and outcomes, contributing to broader research e‌orts aimed at improving the transparency and trustworthiness of automated content analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;subjectivity classication</kwd>
        <kwd>fact-checking</kwd>
        <kwd>misinformation detection</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The CheckThat! lab is organized for the 8th time within CLEF 2025. This paper presents an overview
of Task 1, which covers the challenge of identifying subjectivity in news articles — a task introduced in
the 2023 edition [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and now held for the third time.
      </p>
      <p>As the inuence of digital media has grown, so has the importance of distinguishing between
subjective and objective language. This distinction is paramount in Natural Language Processing
(NLP), especially in domains such as sentiment analysis, opinion mining, and, crucially, fact-checking.
Subjective statements ofien convey personal judgments, emotions, or implicit bias, whereas objective
ones aim to report veriable facts. Automatically recognizing this di‌erence is essential for building
systems that can assess the trustworthiness and neutrality of textual information.</p>
      <p>Task 1 is designed to foster research in this direction by providing a multilingual benchmark for
sentence-level subjectivity classication. Participants are asked to determine whether a sentence taken
from a news article reects the author’s personal viewpoint or o‌ers a neutral, fact-based perspective.
This binary classication task is especially relevant in the current media landscape, where biased
reporting and misinformation pose ongoing challenges to public discourse and information integrity.</p>
      <p>The task includes datasets in nine languages: Arabic, Bulgarian, English, German, Italian, Polish,
Ukrainian, Romanian, and Greek. In particular, the subjectivity task is organized to cover three distinct
settings: monolingual, where the focus is on a specic language; multilingual, where the contribution of
multiple languages is evaluated; and zero-shot, where the generalization capabilities of models trained
on seen languages are tested on unseen ones. All datasets were annotated using a prescriptive framework
designed to support cross-lingual comparability and high annotation quality. System performance
was evaluated using macro-averaged F1 score to ensure a balanced treatment of both subjective and
objective classes. This approach provides a fair and comprehensive measure of system e‌ectiveness
across diverse languages and content.</p>
      <p>The remainder of the paper is as follows. We rst describe the dataset construction process, evaluation
criteria, and submission protocols. We then analyze the submitted systems, comparing their
methodologies and results to assess current progress and identify key challenges. Task 1 contributes to the
broader e‌ort to improve automated understanding of subjectivity in online content — an increasingly
critical component of trustworthy AI applications in the digital era.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Research on subjectivity detection spans a wide array of contexts and has evolved signicantly over
time. While early developments were closely tied to sentiment analysis in English-language texts [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ],
subsequent e‌orts extended to multilingual domains [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], paving the way for cross-lingual approaches.
Over the years, the task has also found relevance in detecting bias [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ], identifying claims [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and
supporting fact-checking workows [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ], which directly motivates the present work.
      </p>
      <p>
        The criteria for identifying subjectivity ofien di‌er depending on the application, and so do the
methodological approaches. Some studies employ lexical heuristics tailored to specic domains or
tasks [
        <xref ref-type="bibr" rid="ref2">2, 11, 12</xref>
        ], while others rely on statistical modeling techniques [13]. A more rigorous path
involves manually curated datasets developed through detailed annotation protocols [14, 15, 16]. As
noted by Chaturvedi et al. [17], these approaches can be grouped into syntactic methods—primarily
rule-based and surface-oriented—and semantic methods, which involve deeper linguistic and contextual
understanding.
      </p>
      <p>Syntactic methods, although ecient in certain settings, typically su‌er from portability issues due
to their dependence on language- or domain-specic indicators. Semantic methods have become more
prevalent as they tend to generalize better, especially when built on systematic annotation schemes. Still,
annotation-driven approaches are not without limitations: disagreements among annotators, vague
or context-sensitive cases, and subjective interpretation introduce inconsistency and noise [15, 18].
Recent work has attempted to mitigate these issues through prescriptive annotation strategies [19],
particularly in the setting of fact verication, where subjective cues are ofien indicative of unveriable
or misleading information [20].</p>
      <p>
        Our work incorporates annotation at multiple textual levels—ranging from isolated sentences [
        <xref ref-type="bibr" rid="ref8">8, 21</xref>
        ],
to text segments [22], and full documents [23]. Although English dominates in terms of available
annotated resources, the eld has seen growing interest in developing datasets for other languages,
including Arabic [24, 25], German [24], French [22], Italian [23], Romanian [26], and Spanish [24].
Nevertheless, many of these e‌orts rely on machine translation and ontology-driven methods for
scalability, which can introduce labeling errors and annotation inconsistencies across languages.
      </p>
      <p>
        This framing has been formalized in recent shared tasks. For instance, the sixth edition of the
CheckThat! lab included a dedicated task on subjectivity detection [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which serves as the foundation
for our current e‌orts. The language coverage has changed slightly since then: due to resource
limitations, Dutch and Turkish were removed, and Bulgarian was added as a new language in the
2024 iteration [27]. In particular, the CheckThat!lab 2024 Task 2 edition [27] also covered multilingual
subjectivity detection, covering ve languages: Arabic, English, German, Italian, and Bulgarian. Our
work builds on this task, extending the set of covered languages to nine, including Polish, Ukrainian,
Romanian and Greek, and exploring zero-shot learning on these unseen languages.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Datasets</title>
      <sec id="sec-3-1">
        <title>3.1. Arabic</title>
        <p>The task o‌ered datasets in nine di‌erent languages with a total of more than 14k sentences manually
annotated following the guidelines in [20]. Table 1 presents details on the dataset statistics. Some
sample instances for each language are given in Table 2.</p>
        <p>For this edition, we used the released dataset from [28] and developed a new test set for the nal
evaluation. The dataset consists of manually annotated sentences from news articles, including sources
such as AraFact [29]. The complete data collection and annotation process involved several phases. In
the article selection phase, 1,159 news articles were selected from AraFact [29]. Additionally, opinionated
articles were manually searched from various Arabic news outlets, resulting in the selection of 221
articles. These articles were parsed and segmented into sentences for annotation.</p>
        <p>The annotation was conducted using the MTurk platform. To ensure annotation quality, standard
qualication tests were applied, and the nal label for each sentence was determined using majority
agreement. A label was assigned to a sentence if at least two annotators agreed. The inter-annotator
agreement (pairwise Cohen’s kappa) was  = 0.538. More details about the data collection and
annotation process can be found in [28].</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. English</title>
        <p>For training, we used NewsSD-ENG [20], a corpus of 1,049 sentences labeled by seven annotators
following guidelines for subjectivity detection tailored to an information retrieval setting [30]. We
merged the dev and dev-test partitions of the CheckThat! lab 2024 Task 2 edition [27] and re-declared
its test split as the new dev-test split. We further collected a novel test set following the same data
collection methodology for NewsSD-ENG. In particular, we retrieved 11 news articles on controversial
topics and randomly sampled 301 sentences. Then, seven annotators labeled the sentences as subjective
or objective. We organized annotators such that each sentence was annotated by three annotators. The
inter-annotator agreement on the new test set measured with Krippendor‌’s alpha was 0.43.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. German</title>
        <p>The German dataset was assembled by randomly selecting sentences from the CT 2022 FAN-Corpus
[31] consisting of news articles that have been annotated according to the factuality of their main claim,
originally. The 800 manually annotated sentences for training and the 491 and 337 instances of the
development and develpment-test sets are from the 2023 and 2024 editions of the task [27]. A new test
set has been annotated following the guidelines outlined in [20]. We excluded all incomplete sentences
as well as non German ones. We also reduced instances consisting of more than one sentence due to
wrong sentence splitting to one sentence. Each sentence has been annotated by the same three native
speakers as in previous iterations of the task, all co-authors of this paper. As the agreement between
the annotators was substantially lower compared to previous years, the annotators discussed every
sentence with deviating labels reaching a consensus (Fleiss’ kappa on the 2025 test set:  = 0.547
( &lt; 0.0001,  = 17.7), Fleiss’ kappa on the 2024 test set:  = 0.696 ( &lt; 0.0001,  = 22.1)).</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Italian</title>
        <p>For training, we used the re-annotated version of SubjectivITA [23] introduced in the CheckThat! lab
2024 Task 2 edition [27]. SubjectivITA is a corpus of news articles annotated for subjectivity detection,
containing 1,841 sentences. We merged the dev and dev-test partitions of the CheckThat! lab 2024
Task 2 edition and re-declared its test split as the new dev-test split. We eventually collected a novel
test split following the same methodology used for the English dataset. In particular, we collected 13
news articles targeting controversial topics and randomly sampled 300 sentences. The inter-annotator
agreement on the new test set measured with Krippendor‌’s alpha [32] was 0.53.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Romanian</title>
        <p>We built the Romanian zero-shot test set from multiple online news websites. In particular, we collected
4 news articles covering controversial topics and randomly sampled 300 instances. Each instance was
labeled as objective or subjective by two native Romanian speakers. The inter-annotator agreement for
the zero-shot Romanian test set, measured using Cohen’s, is 0.30.</p>
        <p>We built the Polish zero-shot test set from multiple online news websites. In particular, we collected 11
news articles covering controversial topics and randomly sampled 350 instances. Each instance was
labeled as objective or subjective by one native Polish speaker.</p>
        <p>We built the Ukrainian zero-shot test set from multiple online news websites. In particular, we collected
17 news articles covering controversial topics and randomly sampled 297 instances. Each instance was
labeled as objective or subjective by one native Ukrainian speaker.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.6. Polish</title>
      </sec>
      <sec id="sec-3-7">
        <title>3.7. Ukrainian 3.8. Greek</title>
        <p>We built the Greek zero-shot test set from multiple online news websites. In particular, we collected 11
news articles covering controversial topics and randomly sampled 300 instances. Each instance was
labeled as objective or subjective by six native Greek speakers. The inter-annotator agreement for the
zero-shot Greek test set, measured using Krippendor‌’s alpha, is 0.36.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Overview of the Systems and Results</title>
      <p>A total of 21 teams participated in the task, submitting 436 valid runs across all language tracks. 16 out of
the 21 teams lled in the survey for the task, providing information about their systems and approaches.
12 teams participated in more than one subtask, while 5 teams opted for only the monolingual English
subtask.</p>
      <p>Table 3 shows the results achieved by the individual teams for each language. Most teams used a
supervised binary classication approach, treating the task as classifying sentences into subjective
(SUBJ) or objective (OBJ). The dominant strategy involved ne-tuning transformer-based models, with
some using ensembles, data augmentation, or additional linguistic features. A few teams explored
probabilistic thresholds, embedding-based classiers, or LLM-based zero-shot and in-context learning
methods. An overview of the approaches is given in Table 4 and a short description of the individual
approaches for each team is given in the following.</p>
      <sec id="sec-4-1">
        <title>4.1. Baselines</title>
        <p>We used the same baseline introduced in the CheckThat! lab 2024 Task 2 edition [27]. In particular, the
baseline was a multilingual SentenceBERT [53] model with a logistic regression classier on top of it. We
considered paraphrase-multilingual-MiniLM-L12-v2 model card as one of the current top-performing
models for semantic similarity. We regularized the logistic regression classier by applying class
reweighting to account for class imbalance. We trained the baseline model on individual language-specic
training data and we evaluated it on the corresponding test set. In the case of zero-shot languages, we
trained the baseline on the multilingual dataset, comprising Arabic, Bulgarian, English, German and
Italian training splits.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results per Language</title>
        <p>Arabic. A total of 12 teams participated in the Arabic subtask, with ve of them not surpassing the
baseline score of 0.5133. Top submissions largely outperformed the baseline, setting a new high score.
In particular, CEA-LIST [37] achieved a macro F1 score of 0.6884 using an ensemble of small language
models and standard encoder-based transformers. The second-ranked team UmuTeam [46] reports
a considerably lower score using MARBERTv2. Likewise Investigators [34], using general-purpose
transformer models like DeBERTa and Multilingual BERT.</p>
        <p>Italian. A total of 13 teams participated in the Italian subtask, with only three of them not surpassing
the baseline score of 0.6941. Team XplaiNLP [43] ranks rst with a F1 score of 0.8104, closely
fdollowed by team CEA-LIST [37]. Team SmolLab_SEU [48] follows with a di‌erence of 3%-points, still
surpassing the baseline by a large margin.</p>
        <p>German. A total of 14 teams participated in the German subtask, with only one team reporting
performance below the baseline score of 0.6960. Team SmolLab_SEU [48] achieved the rst place with
a score of 0.8520. Team UNAM [52] follows with an F1 score of 0.8280. Lastly, team QU-NLP [50]
ranks third with a F1 score of 0.8013. All three top teams largely outperform the baseline with an
improvement of around 10-15%-points.</p>
        <p>English. A total of 22 teams participated in the English subtask, all of them reporting classication
performance above the baseline score of 0.5370. Team QU-NLP [50] ranked rst with a F1 score of
0.8052 using a feature-augmented transformer model. A similar performance is reported by team TIFIN
INDIA [39] with a F1 score of 0.7955. Team CEA-LIST [37] achieves third place with a F1 socre of
0.7739. The large majority of remaining submissions achieved similar results in the range [0.76 - 0.70].
Multilingual. A total of 13 teams participated in the multilingual subtasks, with only three teams
reporting performance below the baseline score of 0.6390. Team TIFIN INDIA [39] ranked rst (0.7550)
with their ensemble of transformer-based models. Teams CEA-LIST [37] and CSECU-Learners [36]
follow with similar classication performance of around ∼ 0.73.</p>
        <p>Polish. A total of 13 teams participated in the Polish subtask, where more than half of the submissions
outperformed the baseline score of 0.5719. In particular, team CEA-LIST [37] ranked rst with a F1
score of 0.6922. Team IIIT Surat [38] reports a ∼ 3-points performance di‌erence using multilingual
BERT. Similarly, Team CSECU-Learners ranks third with a F1 score of 0.6676.</p>
        <p>Ukrainian. A total of 13 teams participated in the Ukrainian subtask. Only four teams managed to
outperform the baseline score of 0.6296, while reporting slightly superior performance. In particular,
team CSECU-Learners [36] achieved rst place with a F1 score of 0.6424. Team Investigators [34]
follows with a F1 score of 0.6413 using a combination of encoder-based models like DeBERTa, BERT,
multilingual BERT and Twitter RoBERTa. Team ClimateSense [40] reports a similar performance to
the top-two teams.</p>
        <p>Romanian. A total of 13 teams participated in the Romanian subtask, with only one team (i.e., TIFIN
INDIA [39]) not surpassing the baseline score of 0.6461. Team QU-NLP [50] ranks rst with a F1
score of 0.8126. There is a ∼ 2-points di‌erence between the rst-ranked team and the second- and
third-ranked teams, namely team CSECU-Learners [36] and team XplaiNLP [43].
Greek. A total of 13 teams participated in the Greek subtask, with around half of the submissions not
surpassing the baseline score of 0.4159. Team AI Wizards [33] ranks rst by ne-tuning a probabilistic
classier on top of DeBERTaV3 model. Similar performance is reported by team SmolLab_SEU [48]
(0.4945) and team CSECU-Learners [36] (0.4919).</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Detailed Description of the Participating Systems</title>
        <p>Below, we describe the approaches of all participating systems; see also Table 4 for an overview.</p>
        <p>Team AI Wizards [33] employed a probabilistic classier with a decision threshold, ne-tuning
DeBERTaV3 for the task.</p>
        <p>Team Investigators [34] utilized encoder-based models including DeBERTa, BERT, Multilingual
BERT, and Twitter RoBERTa.</p>
        <p>Team DSGT-CheckThat [35] ne-tuned encoder models and explored data augmentation
strategies. Their models included RoBERTa (emotion-large), DistilRoBERTa, Sentiment-BERT, ModernBERT,
RoBERTa-large, and MiniLM. They further enhanced performance through Synthetic Data Generation
and Data Augmentation.</p>
        <p>Team CSECU-Learners [36] framed the task as multiclass classication with SUBJ (subjective)
and OBJ (objective) as separate classes. Their transformer models included MPNet, mDeBERTa, and
Multilingual BERT.</p>
        <p>Team CEA-LIST [37] ne-tuned small language models (SLMs) and experimented with LLMs through
techniques such as in-context learning, LLM-as-judge, and model debating. Their models included
RoBERTa, UmBERTo, ALBERTo, Qwen 2.5 70B, Meta-LLaMA 3 70B, DeepSeek 67B, Aya-Expanse-32B,
and GPT-4.1-mini.</p>
        <p>Team IIIT Surat [38] employed a transformer-based model, specically BERT, implemented via
BertForSequenceClassication from Hugging Face, and ne-tuned it for binary classication (SUBJ/OBJ).
They used the pre-trained BERT (English, uncased) for the monolingual classier and Multilingual
BERT (cased) for multilingual and other-language classication, ne-tuning both directly on the CLEF
training data.</p>
        <p>Team TIFIN INDIA [39] used a binary classication approach, where each input is classied as
either subjective or objective. They used an ensemble of transformer-based models and combined their
probability outputs to make the nal prediction post data augmentation. To mitigate data imbalance,
they applied back-translation as a data augmentation technique and used the label distribution ratio to
monitor and address class imbalance. They sed deep learning models based on transformer encoder
architectures, including BERT-Base, BERT-Large, RoBERTa-Base, RoBERTa-Large, XLM-RoBERTa-Base,
XLM-RoBERTa-Large, Modern-BERT-Base, and Modern-BERT-Large. They they applied
probabilitylevel averaging (sofi voting) for model fusion to ensemble predictions across these models. Additionally,
for some datasets, they used a traditional Support Vector Machine (SVM) classier with TF-IDF features
as a lightweight baseline and for comparative analysis. They used a feature-based approach using
Support Vector Machines (SVMs) on selected datasets. The most important features included: TF-IDF
vectors of unigrams and bigrams.</p>
        <p>Team ClimateSense [40] used Embeddings and an MLP classier. They experimented with various
classiers: SVC, Logistic Regression, MLP, etc. They also experimented with various
transformersbased architectures for embedding the sentences: SBERT , RoBERTa-based models, ModernBERT-large,
CT-BERT. Finally, they experimented with Zero-shot prompting some LLMs (such as Zephyr).</p>
        <p>Team CUET_KCRL [41] pursued a supervised classication approach using an LSTM and ne-tuning
mBERT.</p>
        <p>Team nlu@utn [42] followed a Bert-based ensemble model approach, by also adapting the provided
the training data with additional linguistic information before training, using persuasion techniques
identied in the data and POS-counts. The models used were politicalBiasBERT and BERT-base-uncased.</p>
        <p>Team XPlaiNLP [43] employed several transformer-based models, including XLM-RoBERTa-base,
GPT o3-mini, and German-BERT. In particular, for monolingual tasks, German-BERT was ne-tuned on
German and German-translated versions of English, Italian and Bulgarian train datasets.</p>
        <p>Team JU_NLP [54] ne-tuned BERT model on available training data, formulating the task as a
binary classication problem. In particular, they leverage hand-crafied features derived from knowledge
bases and tools like SentiWordNet, WordNet, Opinion lexicon, POS taggers, and lemmatization.</p>
        <p>Team NapierNLP [45] only tackled the English monolingual task by leveraging LLMs. More precisely,
they employed GPT-2, GPTNeo-1.3B, and Qwen3-0.6B. The prompts provided instructions for addressing
the task as a binary classication problem.</p>
        <p>Team UmuTeam [46] employed a wide set of encoder-only transformers, each specic for a given
language. In particular, they employed MARBERTv2 for Arabic data, GottBERT-base for German,
BERTino for Italian, RoBERTa-base for English. Lastly, they used XLM-RoBERTa-base for multilingual
and zero-shot tasks.</p>
        <p>Team UGPLN [47] employed sentence transformers with hand-crafied linguistic features. A logistic
regressor is then trained on top to perform the binary classication task. In particular, they employed
MiniLM-L12-v2 and used the following hand-crafied features: presence of negation cues, sentence
length (i.e., token count), punctuation marks, and lexical opinion indicators derived from the MPQA
Subjectivity lexicon.</p>
        <p>Team SmolLab_SEU [48] employed a vast set of encoder-only transformers, some of which are
language-specic. The models are RoBERTa, DeBERTa-v3, AraBERTv2 and MARBERTv2 for Arabic,
GBERT-large, GottBERT-base, and GElectra-large for German, UmBERTo-v1, and BERT-base-italian for
Italian, MBERT, XLM-RoBERTa-large, InfoXLM-large, MT5-base, and MDeBERTa-v3 for multilingual.
All models were ne-tuned by adding a sequence classication head on top of their pre-trained encoder
layers.</p>
        <p>Team Arcturus [49] ne-tuned the English-pretrained DeBERTa-v3 on monolingual datasets and
evaluate it on all languages, including multilingual and zero-shot tasks.</p>
        <p>Team QU-NLP [50] propose a feature-augmented transformer architecture that combines contextual
embeddings from pre-trained language models with statistical and linguistic features. In particular,
they employed AraElectra for Arabic, augmented with POS tags and TF-IDF features. For cross-lingual
experiments, they employed DeBERTa-v3 with TF-IDF features through a gating mechanism.</p>
        <p>Team CheckMates [51] explored various models such as logistic regression, Support Vector Machine,
BERT, Sentence-BERT, and DistilBERT.</p>
        <p>Team UNAM [52] used di‌erent language-specic versions of the BERT model and focused on
monolingual subtasks.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>We presented an overview of Task 1 from the CheckThat! lab at CLEF 2025. The task concerned the
detection of subjective sentences in controversial news articles. The task was o‌ered in nine di‌erent
languages, four of which were addressed in a zero-shot setting.</p>
      <p>In alignment with the previous edition of the task [27], the majority of the submissions relied on
encoder-only transformer-based architectures, either tailored to a specic language or covering
multilingualism. Some approaches also evaluated popular large language models like GPT with instruction
tuning to detect subjectivity, data augmentation, and automatic translation. The most successful
solutions coupled transformer-based classiers with domain knowledge in the form of feature extraction or
large language models in an ensemble fashion. The best macro F1 scores ranged between 0.50 and 0.85,
showing that annotating and detecting subjectivity present di‌erent challenges that are specic of the
given language. Overall, there is still ample room for improvement in all subtasks. More precisely, in
many cases, we observed that more than half of the teams did not surpass our baseline model.</p>
      <p>As future work, we plan to collect more data concerning existing languages and to expand the set of
covered languages to gather more insights about the task.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>We are thankful to the volunteers that helped with the annotation such Ploutarchos Iliadis, Angeliki
Dimopoulou, Fotini Giannopoulou, Foivos Andrianopoulos, Evangelos Sinos, Panagiotis Reppas who
helped with the annotation of the zero-shot Greek test set.</p>
      <p>The work of F. Alam is partially supported by NPRP 14C-0916-210015 from the Qatar National
Research Fund, part of Qatar Research Development and Innovation Council (QRDI). The work of
J. Struß is partially supported by the BMBF (German Federal Ministry of Education and Research)
under the grant no. 01FP20031J. The work of F. Ruggeri is partially supported by the project European
Commission’s NextGeneration EU programme, PNRR – M4C2 – Investimento 1.3, Partenariato Esteso,
PE00000013 - “FAIR - Future Articial Intelligence Research” – Spoke 8 “Pervasive AI” and by the
European Union’s Justice Programme under Grant Agreement No. 101087342 for the project “Principles
Of Law In National and European VAT”. K. Korre’s research is carried out under the project RACHS:
Rilevazione e Analisi Computazionale dell’Hate Speech in rete, in the framework of the PON programme
FSE REACT-EU, Ref. DOT1303118. A. Muti’s research is supported by the European Research Council
(ERC) under the European Union’s Horizon 2020 research and innovation program (grant agreement
No. 101116095, PERSONAE).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Grammarly for grammar and spelling check.
Afier using this tool/service, the author(s) reviewed and edited the content as needed and take(s) full
responsibility for the publication’s content.
[11] J. Villena-Roma´n, J. Garc´ıa-Morera, M. A´. G. Cumbreras, E. Mart´ınez-Ca´mara, M. T.
Mart´ınValdivia, L. A. U. Lo´pez, Overview of TASS 2015, in: J. Villena-Roma´n, J. Garc´ıa-Morera, M. A´ . G.
Cumbreras, E. Mart´ınez-Ca´mara, M. T. Mart´ın-Valdivia, L. A. U. Lo´pez (Eds.), Proceedings of
TASS 2015: Workshop on Sentiment Analysis at SEPLN co-located with 31st SEPLN Conference
(SEPLN 2015), Alicante, Spain, September 15, 2015, volume 1397 of CEUR Workshop Proceedings,
CEUR-WS.org, 2015, pp. 13–21.
[12] N. Das, S. Sagnika, A subjectivity detection-based approach to sentiment analysis, in: D. Swain, P. K.</p>
      <p>Pattnaik, P. K. Gupta (Eds.), Machine Learning and Information Processing, Springer Singapore,
Singapore, 2020, pp. 149–160.
[13] B. Pang, L. Lee, A sentimental education: Sentiment analysis using subjectivity summarization
based on minimum cuts, in: Proceedings of the 42nd Annual Meeting of the Association for
Computational Linguistics (ACL-04), Barcelona, Spain, 2004, pp. 271–278. doi:10.3115/1218955.
1218990.
[14] J. M. Wiebe, R. F. Bruce, T. P. O’Hara, Development and use of a gold-standard data set for
subjectivity classications, in: Proceedings of the 37th Annual Meeting of the Association for
Computational Linguistics, Association for Computational Linguistics, College Park, Maryland,
USA, 1999, pp. 246–253. doi:10.3115/1034678.1034721.
[15] T. Wilson, J. Wiebe, Annotating opinions in the world press, in: Proceedings of the SIGDIAL 2003
Workshop, The 4th Annual Meeting of the Special Interest Group on Discourse and Dialogue, July
5-6, 2003, Sapporo, Japan, The Association for Computer Linguistics, 2003, pp. 13–22.
[16] M. Abdul-Mageed, M. Diab, Subjectivity and sentiment annotation of Modern Standard Arabic
newswire, in: N. Ide, A. Meyers, S. Pradhan, K. Tomanek (Eds.), Proceedings of the 5th Linguistic
Annotation Workshop, Association for Computational Linguistics, Portland, Oregon, USA, 2011,
pp. 110–118.
[17] I. Chaturvedi, E. Cambria, R. E. Welsch, F. Herrera, Distinguishing between facts and opinions for
sentiment analysis: Survey and challenges, Inf. Fusion 44 (2018) 65–77. doi:10.1016/j.infus.2017.
12.006.
[18] M. Geva, Y. Goldberg, J. Berant, Are we modeling the task or the annotator? An investigation
of annotator bias in natural language understanding datasets, in: K. Inui, J. Jiang, V. Ng, X. Wan
(Eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing
and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019,
Hong Kong, China, November 3-7, 2019, Association for Computational Linguistics, 2019, pp.
1161–1166. doi:10.18653/v1/D19-1107.
[19] P. Ro¨ttger, B. Vidgen, D. Hovy, J. B. Pierrehumbert, Two contrasting data annotation paradigms
for subjective NLP tasks, in: M. Carpuat, M. de Marne‌e, I. V. M. Ru´ız (Eds.), Proceedings
of the 2022 Conference of the North American Chapter of the Association for Computational
Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July
1015, 2022, Association for Computational Linguistics, 2022, pp. 175–190. doi:10.18653/v1/2022.
naacl-main.13.
[20] F. Antici, F. Ruggeri, A. Galassi, K. Korre, A. Muti, A. Bardi, A. Fedotova, A. Barro´n-Cede n˜o, A
corpus for sentence-level subjectivity detection on English news articles, in: N. Calzolari, M.-Y. Kan,
V. Hoste, A. Lenci, S. Sakti, N. Xue (Eds.), Proceedings of the 2024 Joint International Conference
on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), ELRA
and ICCL, Torino, Italia, 2024, pp. 273–285.
[21] S. Rustamov, E. Mustafayev, M. Clements, Sentence-level subjectivity detection using neuro-fuzzy
models, in: A. Balahur, E. van der Goot, A. Montoyo (Eds.), Proceedings of the 4th Workshop on
Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Association for
Computational Linguistics, Atlanta, Georgia, 2013, pp. 108–114.
[22] F. Benamara, B. Chardon, Y. Mathieu, V. Popescu, Towards context-based subjectivity analysis,
in: H. Wang, D. Yarowsky (Eds.), Proceedings of 5th International Joint Conference on Natural
Language Processing, Asian Federation of Natural Language Processing, Chiang Mai, Thailand,
2011, pp. 1180–1188.
[23] F. Antici, L. Bolognini, M. A. Inajetovic, B. Ivasiuk, A. Galassi, F. Ruggeri, SubjectivITA: An
italian corpus for subjectivity detection in newspapers, in: K. S. Candan, B. Ionescu, L. Goeuriot,
B. Larsen, H. Mu¨ller, A. Joly, M. Maistro, F. Piroi, G. Faggioli, N. Ferro (Eds.), Experimental IR
Meets Multilinguality, Multimodality, and Interaction. CLEF 2021, volume 12880 of LNCS, Springer,
2021, pp. 40–52. doi:10.1007/978-3-030-85251-1_4.
[24] C. Banea, R. Mihalcea, J. Wiebe, Multilingual subjectivity: Are more languages better?, in: C.-R.</p>
      <p>Huang, D. Jurafsky (Eds.), Proceedings of the 23rd International Conference on Computational
Linguistics (Coling 2010), Coling 2010 Organizing Committee, Beijing, China, 2010, pp. 28–36.
[25] M. Abdul-Mageed, M. Diab, M. Korayem, Subjectivity and sentiment analysis of Modern Standard
Arabic, in: D. Lin, Y. Matsumoto, R. Mihalcea (Eds.), Proceedings of the 49th Annual Meeting of
the Association for Computational Linguistics: Human Language Technologies, Association for
Computational Linguistics, Portland, Oregon, USA, 2011, pp. 587–591.
[26] C. Banea, R. Mihalcea, J. Wiebe, Sense-level subjectivity in a multilingual setting, Comput. Speech</p>
      <p>Lang. 28 (2014) 7–19. doi:10.1016/j.csl.2013.03.002.
[27] J. M. Struß, F. Ruggeri, A. Barro´n-Ceden˜ o, F. Alam, D. Dimitrov, A. Galassi, G. Pachov, I. Koychev,
P. Nakov, M. Siegel, M. Wiegand, M. Hasanain, R. Suwaileh, W. Zaghouani, Overview of the
CLEF-2024 CheckThat! lab task 2 on subjectivity in news articles, in: [55], 2024.
[28] R. Suwaileh, M. Hasanain, F. Hubail, W. Zaghouani, F. Alam, ThatiAR: Subjectivity detection in
arabic news sentences, arXiv: 2406.05559 (2024).
[29] Z. Sheikh Ali, W. Mansour, T. Elsayed, A. Al-Ali, AraFacts: The rst large Arabic dataset of naturally
occurring claims, in: N. Habash, H. Bouamor, H. Hajj, W. Magdy, W. Zaghouani, F. Bougares,
N. Tomeh, I. Abu Farha, S. Touileb (Eds.), Proceedings of the Sixth Arabic Natural Language
Processing Workshop, Association for Computational Linguistics, Kyiv, Ukraine (Virtual), 2021,
pp. 231–236. URL: https://aclanthology.org/2021.wanlp-1.26.
[30] F. Ruggeri, F. Antici, A. Galassi, K. Korre, A. Muti, A. Barro´n-Cede n˜o, On the denition of
prescriptive annotation guidelines for language-agnostic subjectivity detection, in: R. Campos,
A. M. Jorge, A. Jatowt, S. Bhatia, M. Litvak (Eds.), Text2Story@ECIR, volume 3370 of CEUR
Workshop Proceedings, CEUR-WS.org, 2023, pp. 103–111.
[31] J. Ko¨hler, G. K. Shahi, J. M. Struß, M. Wiegand, M. Siegel, T. Mandl, M. Schu¨tz, Overview of the
CLEF-2022 CheckThat! lab task 3 on fake news detection, in: G. Faggioli, N. Ferro, A. Hanbury,
M. Potthast (Eds.), Working Notes of CLEF 2022—Conference and Labs of the Evaluation Forum,
CLEF ’2022, Bologna, Italy, 2022.
[32] K. Krippendor‌, Computing krippendor‌’s alpha-reliability, 2011. URL: https://repository.upenn.</p>
      <p>edu/handle/20.500.14332/2089.
[33] M. Fasulo, L. Babboni, L. Tedeschini, AI wizards at checkthat! 2025: Enhancing transformer-based
embeddings with sentiment for subjectivity detection in news articles, in: [56], 2025.
[34] S. M. A. Hashmi, S. Aamir, M. Anas, T. Usmani, F. Alvi, A. Samad, Investigators at checkthat! 2025:</p>
      <p>Using llms to improve fact-checking, in: [56], 2025.
[35] M. Heil, D. Bang, DS@GT at checkthat! 2025: Detecting subjectivity via transfer-learning and
corrective data augmentation, in: [56], 2025.
[36] M. Ahmad, A. N. Chy, Csecu-learners at checkthat! 2025: Multilingual transformer-based approach
for subjectivity detection in news articles across multilingual and zero-shot settings, in: [56], 2025.
[37] A. Elbouanani, E. Dufraisse, A. Tuo, A. Popescu, Cea-list at checkthat! 2025: Evaluating llms as
detectors of bias and opinion in text, in: [56], 2025.
[38] S. C. Jaiswal, R. Kumar, Iiit surat at checkthat! 2025: Identifying subjectivity from multilingual
text sequence, in: [56], 2025.
[39] K. Gurumurthy, A. Shrivastava, P. K. Rajpoot, P. Devadiga, B. Hazarika, M. Jain, M. Sharma,
A. Suneesh, A. B. Suresh, A. U. Baliga, Tin at checkthat! 2025: Cross-lingual subjectivity
classication in news through monolingual, multilingual, and zero-shot learning, in: [56], 2025.
[40] G. Burel, P. Lisena, E. Daga, R. Troncy, H. Alani, ClimateSense at CheckThat! 2025: Combining
ne-tuned large language models and conventional machine learning models for subjectivity and
scientic web discourse analysis, in: [56], 2025.
[41] M. T. A. Shawon, F. Haq, M. A. Mia, G. S. M. Mursalin, M. I. Khan, CUET_KCRL at checkthat!2025:</p>
      <p>Ensemblenet with roberta-large for subjectivity detection in news articles, in: [56], 2025.
[42] S. Meyer, M. Roth, nlu@utn at checkthat! 2025: Combining bias sensitivity, linguistic features,
and persuasion cues in an ensemble for subjectivity detection, in: [56], 2025.
[43] A. Sahitaj, J. Li, P. W. Neves, P. S. Fedor Splitt, C. Jakob, V. Solopova, V. Schmitt, XplaiNLP at
checkthat! 2025: Multilingual subjectivity detection with netuned transformers and prompt-based
inference with large language models, in: [56], 2025.
[44] S. Debnath, D. Das, JU_NLP at checkthat! 2025: A condence-guided transformer-based approach
for multilingual subjectivity classication, in: [56], 2025.
[45] K. Alexander, M. Z. Ullah, D. Gkatzia, NapierNLP at checkthat! 2025: Detecting subjectivity with
llms and model fusion, in: [56], 2025.
[46] T. B. Beltra´n, R. Pan, J. A. G. D´ıaz, R. V. Garc´ıa, UmuTeam at checkthat! 2025: Language-specic
versus multilingual models for fact-checking, in: [56], 2025.
[47] M. del Carmen Toapanta-Bernabe´, M. A´ngel Garcia-Cumbreras, L. A. Ure n˜a-Lo´pez, D. D. M.</p>
      <p>Intriago, J. S. Holgu´ın-Reyes, Sinai-UGPLN at checkthat! 2025: A hybrid sbert–logistic regression
framework for segment-level subjectivity detection in english news, in: [56], 2025.
[48] M. A. Rahman, M. A. Amin, M. S. Dewan, M. J. Hasan, M. A. Rahman, SmolLab_SEU at checkthat!
2025: How well do multilingual transformers transfer across news domains for cross-lingual
subjectivity detection?, in: [56], 2025.
[49] A. Aditya, R. Jambulkar, S. Pal, Arcturus at checkthat! 2025: Deberta-v3-base for multilingual
subjectivity detection in news articles, in: [56], 2025.
[50] M. Al-Smadi, Qu-nlp at checkthat! 2025: Multilingual subjectivity in news articles detection using
feature-augmented transformer models with sequential cross-lingual ne-tuning, in: [56], 2025.
[51] R. Padmashri, K. V., V. Srikumar, D. Thenmozhi, Checkmates at checkthat! 2025: Transformer-based
models for subjectivity classication, in: [56], 2025.
[52] I. Diaz, J. Barco, J. Herna´ndez, E. Lee-Romero, G. Bel-Enguix, cepanca_UNAM at checkthat! 2025:</p>
      <p>Using bert-based classiers for detection of subjectivity, in: [56], 2025.
[53] N. Reimers, I. Gurevych, Sentence-bert: Sentence embeddings using siamese bert-networks,
in: K. Inui, J. Jiang, V. Ng, X. Wan (Eds.), Proceedings of the 2019 Conference on Empirical
Methods in Natural Language Processing and the 9th International Joint Conference on Natural
Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, Association
for Computational Linguistics, 2019, pp. 3980–3990. doi:10.18653/V1/D19-1410.
[54] A. A. M. Sardar, K. Fatema, M. A. Islam, JUNLP at CheckThat! 2024: Enhancing check-worthiness
and subjectivity detection through model optimization, in: [55], 2024.
[55] G. Faggioli, N. Ferro, P. Galusˇcˇa´kova´, A. Garc´ıa Seco de Herrera (Eds.), Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, CLEF 2024, 2024.
[56] G. Faggioli, N. Ferro, P. Rosso, D. Spina (Eds.), Working Notes of CLEF 2025 - Conference and Labs
of the Evaluation Forum, CLEF 2025, Madrid, Spain, 2025.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barro´</surname>
          </string-name>
          n-Cede n˜o,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Struss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Antici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          , J. Ko¨hler,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Leistra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Mehmet</given-names>
            <surname>Deniz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          , W. Zaghouani,
          <article-title>Overview of the CLEF-2023 CheckThat! lab: Task 2 on subjectivity in news articles</article-title>
          , in: M.
          <string-name>
            <surname>Aliannejadi</surname>
            , G. Faggioli,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          , Vlachos, Michalis (Eds.), Working Notes of CLEF 2023-
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2023</year>
          , Thessaloniki, Greece,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          , E. Rilo‌,
          <article-title>Creating subjective and objective sentence classiers from unannotated texts</article-title>
          , in: A. F. Gelbukh (Ed.),
          <source>Computational Linguistics and Intelligent Text Processing</source>
          , volume
          <volume>3406</volume>
          , Springer, Berlin and Heidelberg,
          <year>2005</year>
          , pp.
          <fpage>486</fpage>
          -
          <lpage>497</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>540</fpage>
          -30586-6_
          <fpage>53</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wilson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ho</surname>
          </string-name>
          <article-title>‌mann, Recognizing contextual polarity in phrase-level sentiment analysis</article-title>
          , in: R.
          <string-name>
            <surname>Mooney</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Brew</surname>
            ,
            <given-names>L.-F.</given-names>
          </string-name>
          <string-name>
            <surname>Chien</surname>
          </string-name>
          , K. Kirchho‌ (Eds.),
          <source>Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Morristown, NJ and USA,
          <year>2005</year>
          , pp.
          <fpage>347</fpage>
          -
          <lpage>354</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Banea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          ,
          <article-title>Learning multilingual subjective language via cross-lingual projections</article-title>
          , in: A.
          <string-name>
            <surname>Zaenen</surname>
          </string-name>
          , A. van den Bosch (Eds.),
          <source>ACL</source>
          <year>2007</year>
          ,
          <article-title>Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics</article-title>
          , June 23-30,
          <year>2007</year>
          , Prague, Czech Republic, The Association for Computer Linguistics,
          <year>2007</year>
          , pp.
          <fpage>976</fpage>
          -
          <lpage>983</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Banea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <article-title>Multilingual subjectivity analysis using machine translation</article-title>
          , in: M.
          <string-name>
            <surname>Lapata</surname>
          </string-name>
          , H. T. Ng (Eds.),
          <source>Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Stroudsburg, PA and USA,
          <year>2008</year>
          , pp.
          <fpage>127</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Aleksandrova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lareau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.-A.</given-names>
            <surname>Me</surname>
          </string-name>
          <article-title>´nard, Multilingual sentence-level bias detection in wikipedia</article-title>
          , in: R. Mitkov, G. Angelova (Eds.),
          <source>Proceedings of the International Conference on Recent Advances in Natural Language Processing</source>
          , Incoma Ltd.,
          <string-name>
            <surname>Shoumen</surname>
          </string-name>
          , Bulgaria,
          <year>2019</year>
          , pp.
          <fpage>42</fpage>
          -
          <lpage>51</lpage>
          . doi:
          <volume>10</volume>
          .26615/
          <fpage>978</fpage>
          -954-452-056-4_
          <fpage>006</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C.</given-names>
            <surname>Hube</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fetahu</surname>
          </string-name>
          ,
          <article-title>Neural based statement classication for biased language</article-title>
          ,
          <source>in: Proceedings of the Twelfih ACM International Conference on Web Search and Data Mining</source>
          , Association for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , pp.
          <fpage>195</fpage>
          -
          <lpage>203</lpage>
          . doi:
          <volume>10</volume>
          .1145/3289600.3291018.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>E. Rilo‌</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiebe</surname>
          </string-name>
          ,
          <article-title>Learning extraction patterns for subjective expressions</article-title>
          ,
          <source>in: Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2003</year>
          , pp.
          <fpage>105</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L. L.</given-names>
            <surname>Vieira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L. M.</given-names>
            <surname>Jeronimo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. E. C.</given-names>
            <surname>Campelo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. B.</given-names>
            <surname>Marinho</surname>
          </string-name>
          ,
          <article-title>Analysis of the subjectivity level in fake news fragments</article-title>
          ,
          <source>in: Proceedings of the Brazilian Symposium on Multimedia and the Web</source>
          ,
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <year>2020</year>
          , pp.
          <fpage>233</fpage>
          -
          <lpage>240</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 3428658.3430978.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>C. L. M. Jeroˆnimo</surname>
          </string-name>
          , L. B.
          <string-name>
            <surname>Marinho</surname>
            ,
            <given-names>C. E. C.</given-names>
          </string-name>
          <string-name>
            <surname>Campelo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Veloso</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Da Costa</surname>
            <given-names>Melo</given-names>
          </string-name>
          ,
          <source>Fake News Classication Based on Subjective Language</source>
          , in: M.
          <string-name>
            <surname>Indrawan-Santiago</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Pardede</surname>
            ,
            <given-names>I. L.</given-names>
          </string-name>
          <string-name>
            <surname>Salvadori</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Steinbauer</surname>
            , I. Khalil,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Anderst-Kotsis</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 21st International Conference on Information Integration and Web-Based Applications &amp; Services</source>
          , iiWAS2019, Association for Computing Machinery, New York, NY, USA,
          <year>2020</year>
          , pp.
          <fpage>15</fpage>
          -
          <lpage>24</lpage>
          . doi:
          <volume>10</volume>
          .1145/3366030.3366039.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>