<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>HYBRINFOX at CheckThat! 2024 - Task 2: Enriching BERT Models with the Expert System VAGO for Subjectivity Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Morgane Casanova</string-name>
          <email>morgane.casanova@irisa.fr</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julien Chanson</string-name>
          <email>julien.chanson@mondeca.com</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benjamin Icard</string-name>
          <email>benjamin.icard@lip6.fr</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Géraud Faye</string-name>
          <email>geraud.faye@centralesupelec.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guillaume Gadek</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guillaume Gravier</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Paul Égré</string-name>
          <email>paul.egre@ens.psl.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Airbus Defence and Space</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institut Jean-Nicod</institution>
          ,
          <addr-line>CNRS, ENS-PSL, EHESS</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>LIP6, Sorbonne Université</institution>
          ,
          <addr-line>CNRS</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Mondeca</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Université Paris-Saclay</institution>
          ,
          <addr-line>CentraleSupélec, MICS</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Université de Rennes</institution>
          ,
          <addr-line>CNRS, Inria, IRISA</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents the HYBRINFOX method used to solve Task 2 of Subjectivity detection of the CLEF 2024 CheckThat! competition. The specificity of the method is to use a hybrid system, combining a RoBERTa model, ifne-tuned for subjectivity detection, a frozen sentence- BERT (sBERT) model to capture semantics, and several scores calculated by the English version of the expert system VAGO, developed independently of this task to measure vagueness and subjectivity in texts based on the lexicon. In English, the HYBRINFOX method ranked 1st with a macro F1 score of 0.7442 on the evaluation data. For the other languages, the method used a translation step into English, producing more mixed results (ranking 1st in Multilingual and 2nd in Italian over the baseline, but under the baseline in Bulgarian, German, and Arabic). We explain the principles of our hybrid approach, and outline ways in which the method could be improved for other languages besides English.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Subjectivity</kwd>
        <kwd>Objectivity</kwd>
        <kwd>Vagueness</kwd>
        <kwd>Hybrid AI</kwd>
        <kwd>VAGO</kwd>
        <kwd>Large Language Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Detecting subjectivity in natural language is an important task for a number of applications in the
domain of news and communication, with strong ties to disinformation and propaganda that constitute
the playground of the HYBRINFOX project. Indeed, objective statements in news can be defined as
statements that are open to verification by limiting bias and interpretive disagreement. While an
objective statement may turn out false, the information it conveys is generally taken to be trustworthy,
because it is prone to independent confirmation and fact-checking.</p>
      <p>On the contrary, subjective statements convey personal feelings and opinions. By definition, they
are prone to inter-personal disagreement and do not obey the same norms of justification and
verification. While subjectivity may appear in explicit opinion papers that are not necessarily considered
as propaganda or as manipulation, it is also widely used implicitly in conjunction with false objective
statements, or just to bias true objective information.</p>
      <p>
        Task 2 of the CLEF 2024 CheckThat! benchmark [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ], running for the second time since
2023, aimed at detecting subjective utterances and thus met the objectives of the HYBRINFOX project
that seek to develop hybrid methods for the identification of vague information likely to introduce
or encourage bias (subjectivity, evaluativity). In particular, HYBRINFOX aims to develop tools for
measuring linguistic vagueness in texts, taking advantage of a symbolic AI method, VAGO, to improve
the performance of deep learning models [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. The project also explores the boundary between truthful
and untruthful uses of linguistic vagueness in discourse [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
      <p>
        Subjectivity, vagueness, uncertainty, speculations, expressions of opinions are rather dificult to
detect automatically in texts as these concepts involve several pragmatic and rhetorical aspects that go
beyond the lexical semantics for which most NLP models are designed. The dificulty also translates in
the definition of subjectivity, where annotation guidelines have evolved over time – see, e.g., [
        <xref ref-type="bibr" rid="ref10 ref2 ref9">9, 10, 2</xref>
        ].
There is, however, a vast literature on these topics, with a number of recent contributions as part of
previous editions of the CheckThat! benchmark [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        The HYBRINFOX method that competed explores the augmentation of an LLM-based system with
vagueness scores given by an expert system primarily based on lexical analysis as illustrated in Figure 1.
The method in fact combines a RoBERTa model, fine-tuned for subjectivity detection, a frozen
sentenceBERT (sBERT) model to capture semantics, and the scores calculated by the English version of the expert
system VAGO [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].1 Dimensionality reduction operates on the concatenation of the BERT and
sentenceBERT embeddings before combining with the vagueness scores as input to a classification linear layer.
We justify the selection of this package by comparison with other, less eficient combinations, on the
developmental data (see Table 2 below).
      </p>
      <p>The hybrid system was primarily developed for English, and then adapted to other languages, relying
on automatic translation into English for the other test languages. The method ranked 1st in English on
the evaluation data (with a macro F1 score of 0.7442), but in other languages it produced more mixed
results (ranking 1st on Multilingual and 2nd in Italian over the baseline, but 3rd in Bulgarian, 4th in
German, and 6th in Arabic under the baseline), likely due to loss in accuracy caused by the translation
step.</p>
      <p>
        The paper is organized as follows. We start with a brief state of the art in Section 2. In Section 3 we
introduce the expert system VAGO designed to produce vagueness and subjectivity scores. Section 4
reports on experiments conducted towards the choice of an adequate hybrid system, and presents
comparative results on the development set. Section 5 analyzes the results obtained in the evaluation
phase, including post hoc analyses. Finally, Section 6 concludes with suggestions on how to extend and
improve the current results.
1In Checkthat! 2024 Task 1, our team explored a distinct but similarly hybrid approach for checkworthiness estimation, see
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] for details.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. State of the art</title>
      <p>
        Subjectivity and vagueness detection has been addressed with expert systems, often relying on lexical
analysis as well as on patterns, viz. [
        <xref ref-type="bibr" rid="ref12 ref14">14, 15, 12</xref>
        ]. To some extent, [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] also makes use of linguistic heuristics
to determine the uncertainty scope within an utterance. More recent work, however, relies on statistical
models, such as [16, 17], and all systems at the 2023 benchmark were based on large language models
(LLMs) such as BERT or GPT [18, 19, 20, 21, 22, 23, 24]. Most systems explored fine-tuning diferent
language models for subjectivity detection, either in a multilingual or in a language-specific setting.
Some systems added components to cope with data imbalance and scarcity, here again leveraging large
language models including generative ones for data augmentation [18, 20]. While statistical systems
do perform well, they lack explicit features of expert systems that make them explainable and that we
believe can make them more eficient. Conversely, expert systems make very limited use of contextual
features. The gist of our approach, therefore, is to define a method drawing on both approaches and
combining their respective strengths.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Symbolic scoring using VAGO</title>
      <p>VAGO relies on a symbolic approach to assign scores of vagueness, subjectivity, detail, and objectivity
to sentences. Regarding the measure of detail and objectivity of a text, the detection is based in part on
identifying named entities (including people, locations, temporal indications, institutions, and numbers),
using the open-source library for Natural Language Processing spaCy.2 Our underlying assumption is
that such entities ground the information reported in specific objects and generally leave very limited
room for variable interpretation. The more named entities, the more detailed the sentence is likely to
be. Given a sentence , we say that its category is   if it names an entity. The class   is not closed,
as its members are determined by named entity recognition.</p>
      <p>
        By contrast, the detection of vagueness and subjectivity relies on a closed but evolving lexical database,
which consisted of 1,614 terms in English at the time of the CheckThat! 2024 experiment [25]. Derived
from a typology of vagueness proposed in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], this database provides an inventory of lexical items
distributed in four categories for vagueness (approximation vagueness (), generality vagueness (),
degree vagueness (), combinatorial vagueness ( ), and an additional category of explicit markers
of subjectivity (), not counted as vague.
      </p>
      <p>Expressions of approximation vagueness include modifiers like “approximately”, which relax the
truth conditions of the modified expression. Generality vagueness includes determiners like “some” and
modifiers such as “at most”. The category of expressions related to degree vagueness and combinatorial
vagueness [26] mainly consists of one-dimensional gradable adjectives (such as “tall” and “old”) and
multidimensional gradable adjectives, including a number of evaluative adjectives (like “beautiful”,
“intelligent”, “good”, or “skilled”). These expressions, unlike expressions of generality vagueness, lack
precise truth-conditions, and they leave room for inter-personal disagreement and subjectivity [27, 28,
29, 30]. Finally, explicit markers of subjectivity include separate expressions such as exclamation marks
(“!”), first-person pronouns (“I/we”), and some expressive adverbs (“ever”, “of course”). As seen in Table
1, the category  is much bigger than the others, and as we will see below it also plays a determinant
role in the detection of subjectivity.</p>
      <p>Expressions of type  , , , are treated as factual or objective, whereas expressions of type ,
,  , are treated as subjective. Table 1 provides the exact numbers of items per category available in
the English VAGO database used for the CheckThat! 2024 competition (version of May 2024).</p>
      <p>For each measurement, relevant markers are detected and scored from words to sentences. For a
given sentence , its vagueness score is calculated as the ratio between the sum of vague words in the
sentence, | |, and the total number of words in the sentence, notated . That is:
subjective objective
() = || +⏟ | | + || + || = | |
⏞ ⏞ ⏟
 
|| + || + | |:
where ||, ||, ||, and | | represent the number of terms in  belonging to each of the four
vagueness categories.</p>
      <p>The subjectivity score of a sentence  is calculated as the ratio between the subjective expressions
in , either vague or explicit markers, and the total number of words in . That is, letting || =</p>
      <p>The detail-vs-vagueness score of a sentence  is defined as the relative proportion of named entities
| | in the sentence, compared to the number of vague terms in  (across all categories) | |:
() = ||</p>
      <p>/() =</p>
      <p>| |
| | + | |</p>
      <p>The objectivity-vs-subjectivity score of a sentence is defined as the relative proportion of objective
expressions in  (objective vagueness and named entities), compared to the number of subjective terms
in . That is, letting || = | | + || + ||:
/() =</p>
      <p>||
|| + ||</p>
      <p>In summary, the symbolic method encodes objectivity in terms of two main dimensions: named
entities ( ) and objective vague expressions. And likewise it encodes subjectivity in terms of subjective
vague expressions and explicit markers of subjectivity. Hence, vagueness and subjectivity overlap, but
neither includes the other, and vagueness does not rule out objectivity. Expressions of detail are all
markers of objectivity, but objectivity is a broader category.</p>
      <p>
        VAGO does not consist solely of a lexicon and scoring rules, but it also includes expert rules of
vagueness-cancellation (viz. the measure phrase “180cm” in “Mary is 180cm tall” cancels the vagueness
and subjectivity of the adjective “tall” when it occurs unmodified as in “Mary is tall”). Quotation marks
are also handled as cancelling the subjectivity of terms occurring within the marks. This choice agrees
with with the annotation guide proposed in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], in which reported speech is taken to be conveying
objective information on views that may themselves be subjective.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Development phase: defining an optimal hybrid system</title>
      <p>During the development phase, we tested and compared six variants of our system on the English data.
Two variants are purely machine-learning based and serve as a baseline to measure the contribution of
VAGO to a hybrid system. The other four variants address diferent ways of integrating VAGO features
to a machine-learning approach. Each system was fine-tuned on the English training data over 30
epochs, with a batch size of 6 and a learning rate of 10e-6. Results for the diferent systems are reported
in Table 2. For languages other than English, we used the English system on an initial translation into
English using the DeepL translator3. We thus only report results on English for the development phase.</p>
      <p>As an initial baseline, a first pure machine-learning system consists in fine-tuning a BERT model
for document classification, specifically the “ RoBERTa-base” model, with a single value output 1 for
subjective, and 0 for objective. We chose RoBERTa based on its good performances on text classification
and for reasons of familiarity, leaving open whether other models could be used instead.</p>
      <p>The optimization criterion at training is the binary cross-entropy. At test time, we compare the
score given by the system to a threshold, utterances with a score above the threshold being deemed as
subjective.4 By varying the threshold, we obtain diferent trade-ofs in terms of misses and false alarms
for subjective utterances, yielding ROC curves as reported in Figure 2, where the red curve corresponds
to the baseline. We defined the optimal threshold on the test set of the development data, searching for
the threshold maximizing the oficial macro F1 metric. Optimal thresholds found on the validation data
along with the corresponding scores are reported in Table 2, where the ▷ symbol shows the system
retained.</p>
      <p>Due to the fine-tuning of all the parameters, the resulting model is likely to encode mostly lexical
information dedicated to the task at hand, disregarding semantics. We thus tested a variant combining
the BERT model with a sentence-BERT model whose parameters are frozen. The sentence-BERT model
is a BERT model trained within a siamese architecture to yield sentence embeddings that are close
one to another in the embedded space for utterances having similar meanings. These models, trained
on paraphrasing and on natural language inference tasks, are known to encode semantics to some
extent, thus providing our system with an intuition of the meaning to facilitate classification. In the
experiments, we used the “distilbert-base-nli-mean-tokens” checkpoint and did not re-estimate the
parameters at training time to keep general semantics. Adding a semantic description as input to the
classifier improved results as reported in Table 2.</p>
      <p>Making use of the expert system VAGO in a BERT-based approach can rely on one of two features:
(i) the terms detected by VAGO for the five categories ( , , ,  , ) in an utterance, referred to
as “VAGO Terms”; (ii) the four vagueness scores described in the previous section, referred to as “VAGO
Scores”.</p>
      <p>In the first case, a straightforward approach simply consists in augmenting the input utterance with
the VAGO terms before training a BERT classifier. The input to the BERT classifier thus consists of the
utterance to classify, followed by a list of VAGO terms separated from the utterance with a [SEP] token.
3https://www.deepl.com/
4Note that training is thus done with an implicit threshold equal to 0.5.</p>
      <p>The idea of this approach, designated as “RoBERTa + VAGO Terms”, is to reinforce the decision by
emphasizing the terms deemed of interest by VAGO. In the case of VAGO scores, the idea is to combine
the BERT utterance embedding with the VAGO scores before the classification. To prevent the four
VAGO scores from being overwhelmed by the 768 dimensions of the BERT utterance embedding, we
ifrst reduced the dimension of the latter to 5 before concatenation with the VAGO scores. The resulting
9 dimensional feature vector constitutes the input to the classification head. Both strategies improve
performance over the BERT baseline, with VAGO scores being slightly more eficient than VAGO terms.</p>
      <p>We also combined these last two hybrid approaches with the sentence-BERT embeddings as discussed
previously. The combination of BERT and sentence-BERT embeddings with the VAGO Scores defines
the oficial HYBRINFOX system that competed in CLEF, achieving the best results on the development
set. It is illustrated in Figure 1 and marked with the symbol ▷ in Table 2. The BERT and
sentenceBERT embeddings are concatenated before reducing the 2x768 dimensional vector to 5 with a linear
projection. As previously, this last feature vector is concatenated with the VAGO Scores before entering
classification layer. We also tested the addition of the VAGO Terms to the BERT encoder in this last
architecture, however with no success, yielding a reduction of the subjective F1 score. Interestingly, the
ROC curves show a strong benefit for the “ RoBERTa + sBERT + VAGO Scores” over the baseline, where
a significant improvement of the false positive rate of 0.25 is observed for a fixed true positive rate of
90 %.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation phase: results and analyses</title>
      <p>and annotations.</p>
      <p>We also conducted a post-evaluation analysis of the optimal decision threshold, which was empirically
set at 0.1 on the development data. Results are reported in the right-most columns of Table 3. Overall,
the threshold of 0.1 appears to be stable across languages, however it is suboptimal for German, Arabic
and for the Multilingual dataset. For these three datasets, better performance would have been achieved
by lowering the threshold to 0.05. In other words, with a threshold of 0.1, the miss rate for subjective
utterances was probably too high to yield a good compromise between recall and precision for an
optimal macro F1 score. This observation calls for further work on score normalization towards better
stability across languages and datasets.</p>
      <p>Upon further analysis, we discovered that our results for Bulgarian could have been improved through
additional corpus cleaning. Specifically, square bracket symbols were inadvertently included in the
translations into English, which we did not address during the evaluation phase. By removing these
brackets, our results for Bulgarian improved significantly, achieving a Macro F1 score of 0.7561 and
a SUBJ F1 score of 0.7122 with an optimal threshold of 0.1. This issue with square brackets was also
present in other languages, but did not afect the results as significantly as it did for Bulgarian.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and future work</title>
      <p>To solve Task 2 of Subjectivity Detection, we used a hybrid approach combining the rigid symbolic AI
system VAGO rules with flexible BERT predictions taking context into account. Our main target was
English, one of the languages of VAGO, for which we obtained good results in the development and
in the evaluation phases. For other languages, we used the English system on automatic translations
obtained by DeepL.</p>
      <p>In order to improve our approach, we believe that separate VAGO lexicons should be developed for
additional languages besides English and French, in order to get rid of the translation step into English,
and to be able to use appropriate BERT and sBERT models for each language. The development of an
Italian lexicon was under way at the time of the evaluation, but not suficiently advanced yet to produce
satisfactory results. Alternatively, we could aim for a better control of the quality of the translation into
English, our assumption being that the lexicon of objectivity and subjectivity ought to be preserved
under translation.</p>
      <p>
        Finally, we stress that our approach of subjectivity detection was developed independently of the
task. During the training phase, we noticed that some expressions that VAGO classifies as subjective,
like the vague determiner “many”, were not systematically associated with the label “subjective” (31 out
of 42 sentences including “many” are labelled as objective in the development set provided ahead of the
evaluation phase). We did not try to modify VAGO’s classification principles on this or other specific
entries. Instead, we left it in place in order to see better if our understanding of subjectivity and the
understanding of subjectivity proposed in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] would agree. We were pleased to see that they mostly do,
because this lends support to the hypothesis that the expression of subjectivity is characterized in part
by the use of a specific lexicon and by specific rhetorical markers.
      </p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>We thank two anonymous referees for helpful comments. This work was supported by the programs
HYBRINFOX (ANR-21-ASIA-0003), FRONTCOG (ANR-17-EURE-0017), and THEMIS (n°DOS0222794/00
and n° DOS0222795/00). PE thanks Monash University for hosting him during the writing of this paper,
in the context of the program PLEXUS (Marie Skłodowska-Curie Action, Horizon Europe Research and
Innovation Programme, grant n°101086295).
detection of uncertainty in natural language requirements, in: IEEE International Requirements
Engineering Conference, 2012, pp. 11–20.
[15] E. Rilof, J. Wiebe, Learning extraction patterns for subjective expressions, in: Conference on</p>
      <p>Empirical Methods in Natural Language Processing, 2003, pp. 105–112.
[16] H. Huo, M. Iwaihara, Utilizing bert pretrained models with various fine-tune methods for
subjectivity detection, in: Asia-Pacific Web and Web-Age Information Management Joint International
Conference on Web and Big Data, 2020, pp. 270–284.
[17] S. Sagnika, B. S. P. Mishra, S. K. Meher, An attention-based CNN-LSTM model for subjectivity
detection in opinion-mining, Neural Computing and Applications 33 (2021) 17425–17438.
[18] I. Baris Schlicht, L. Khellaf, D. Altiok, DWReCO at CheckThat! 2023: Enhancing subjectivity
detection through style-based data sampling, in: Working Notes of the Conference and Labs of
the Evaluation Forum, 2023, pp. 306–317.
[19] K. Dey, P. Tarannum, M. A. Hasan, S. R. H. Noori, NN at CheckThat! 2023: Subjectivity in news
articles classification with transformer based models, in: Working Notes of the Conference and
Labs of the Evaluation Forum, 2023, pp. 318–328.
[20] R. A. Frick, Fraunhofer sit at CheckThat! 2023: can LLMs be used for data augmentation &amp;
few-shot classification? detecting subjectivity in text using chatGPT, in: Working Notes of the
Conference and Labs of the Evaluation Forum, 2023, pp. 329–336.
[21] F. A. Leistra, T. Caselli, Thesis Titan at CheckThat! 2023: Language-specific fine-tuning of
mDeBERTaV3 for subjectivity detection, in: Working Notes of the Conference and Labs of the
Evaluation Forum, 2023, pp. 351–359.
[22] G. Pachov, D. Dimitrov, I. Koychev, P. Nakov, Gpachov at CheckThat! 2023: a diverse
multiapproach ensemble for subjectivity detection in news articles, in: Working Notes of the Conference
and Labs of the Evaluation Forum, 2023, pp. 404=–412.
[23] H. T. Sadouk, F. Sebbak, H. E. Zekiri, ES-VRAI at CheckThat! 2023: Enhancing model performance
for subjectivity detection through multilingual data aggregation, in: Working Notes of the
Conference and Labs of the Evaluation Forum, 2023, pp. 423–429.
[24] S. Tran, P. Rodrigues, B. Strauss, E. Williams, Accenture at CheckThat! 2023: Impacts of
backtranslation on subjectivity detection, in: Working Notes of the Conference and Labs of the
Evaluation Forum, 2023, pp. 507–517.
[25] G. Atemezing, B. Icard, P. Égré, Vague Terms in SKOS to detect vagueness in textual documents,
2022. URL: https://doi.org/10.5281/zenodo.4718530. doi:10.5281/zenodo.4718530.
[26] W. P. Alston, Philosophy of Language, Prentice Hall, 1964.
[27] C. Wright, The epistemic conception of vagueness, The Southern journal of philosophy 33 (1995)
133.
[28] C. Kennedy, Two sources of subjectivity: Qualitative assessment and dimensional uncertainty,</p>
      <p>Inquiry 56 (2013) 258–277.
[29] S. Verheyen, S. Dewil, P. Égré, Subjectivity in gradable adjectives: The case of tall and heavy,</p>
      <p>Mind &amp; Language 33 (2018) 460–479.
[30] S. Solt, Multidimensionality, subjectivity and scales: Experimental evidence, in: E. Castroviejo,
L. McNally, G. Sassoon (Eds.), The Semantics of Gradability, Vagueness, and Scale Structure,
Springer, 2018, pp. 59–91.
[31] G. Faggioli, N. Ferro, P. Galuščáková, A. García Seco de Herrera (Eds.), Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, CLEF 2024, Grenoble, France, 2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Antici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Köhler</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the CLEF-2023 CheckThat! Lab: Task 2 on subjectivity in news articles</article-title>
          ,
          <source>in: Working Notes of the Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>249</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Antici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Korre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <article-title>On the definition of prescriptive annotation guidelines for language-agnostic subjectivity detection</article-title>
          .,
          <source>Text2Story @ European Coference on Information Retrieval</source>
          <volume>3370</volume>
          (
          <year>2023</year>
          )
          <fpage>103</fpage>
          -
          <lpage>111</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Przybyła</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <surname>The</surname>
            <given-names>CLEF</given-names>
          </string-name>
          -2024 CheckThat! Lab:
          <article-title>Check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>449</fpage>
          -
          <lpage>458</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Wiegand, Overview of the CLEF-2024 CheckThat! Lab Task 2 on subjectivity in news articles</article-title>
          ,
          <source>in: [31]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Guélorget</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          , G. Gadek,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gahbiche</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gatepaille</surname>
          </string-name>
          , G. Atemezing, P. Égré,
          <article-title>Combining vagueness detection with deep learning to identify fake news</article-title>
          ,
          <source>in: IEEE International Conference on Information Fusion</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Claveau</surname>
          </string-name>
          , G. Atemezing,
          <string-name>
            <given-names>P.</given-names>
            <surname>Égré</surname>
          </string-name>
          ,
          <article-title>Measuring vagueness and subjectivity in texts: from symbolic to neural VAGO</article-title>
          ,
          <source>in: IEEE International Conference on Web Intelligence and Intelligent Agent Technology</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>395</fpage>
          -
          <lpage>401</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Égré</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          ,
          <article-title>Lying and vagueness</article-title>
          , in: J.
          <string-name>
            <surname>Meibauer</surname>
          </string-name>
          (Ed.), Oxford Handbook of Lying, OUP,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Égré</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Spector</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mortier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Verheyen</surname>
          </string-name>
          ,
          <article-title>On the optimality of vagueness: “around”, “between” and the Gricean maxims</article-title>
          ,
          <source>Linguistics and Philosophy</source>
          <volume>46</volume>
          (
          <year>2023</year>
          )
          <fpage>1075</fpage>
          -
          <lpage>1130</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wilson</surname>
          </string-name>
          , J. Wiebe, Annotating opinions in the world press,
          <source>in: Proceedings of the Fourth SIGdial Workshop of Discourse and Dialogue</source>
          ,
          <year>2003</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Kennedy</surname>
          </string-name>
          ,
          <article-title>Two kinds of subjectivity</article-title>
          , in: C.
          <string-name>
            <surname>Meier</surname>
          </string-name>
          , J. van Wijnbergen-Huitink (Eds.), Subjective meaning: Alternatives to relativism, Walter de Gruyter GmbH &amp;
          <string-name>
            <surname>Co</surname>
            <given-names>KG</given-names>
          </string-name>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galassi</surname>
          </string-name>
          , G. Da San Martino, P. Nakov,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Azizov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. S.</given-names>
            <surname>Cheema</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Haouari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ruggeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Struß</surname>
          </string-name>
          , W. Zaghouani,
          <article-title>Overview of the CLEF-2023 CheckThat! Lab on checkworthiness, subjectivity, political bias, factuality, and authority of news articles and their source</article-title>
          ,
          <source>in: Working Notes of the Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2023</year>
          , p.
          <fpage>251</fpage>
          -
          <lpage>275</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          , G. Atemezing, P. Égré,
          <article-title>VAGO: un outil en ligne de mesure du vague et de la subjectivité</article-title>
          , in: Conférence Nationale sur les Applications Pratiques de l'
          <source>Intelligence Artificielle</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>68</fpage>
          -
          <lpage>71</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Faye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Casanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Icard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chanson</surname>
          </string-name>
          , G. Gadek, G. Gravier, P. Égré, HYBRINFOX at CheckThat! 2024:
          <article-title>Enhancing language models with structured information for checkworthiness estimation</article-title>
          ,
          <source>in: [31]</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. De Roeck</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Gervasi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Willis</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Nuseibeh</surname>
          </string-name>
          , Speculative requirements: Automatic
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>