<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Eevvgg at CheckThat! 2024: Evaluative Terms, Pronouns and Modal Verbs as Markers of Subjectivity in Text</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ewelina Gajewska</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Warsaw University of Technology</institution>
          ,
          <addr-line>Plac Politechniki 1, 00-661 Warsaw</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>This work tests performance of simple machine learning algorithms against large language models (LLMs) utilising transfer learning for a binary detection of subjectivity in news articles. Second, the influence of feature normalisation on classification performance is examined. Third, the work measures impact of training data size on subjectivity extraction. The proposed BERTd model that makes use of additional information about stance markers in news articles was placed 8th in the oficial ranking of the CLEF 2024 CheckThat! lab Task 2: Subjectivity in News Articles competition for English data, achieving 0.70 macro-averaged 1. Models that distinguish subjective opinions from objective facts could be utilised in studies on information verification (detection of fake news, understood as a mixture of subjective opinion and facts).</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;stance</kwd>
        <kwd>subjectivity</kwd>
        <kwd>fake news</kwd>
        <kwd>text classification</kwd>
        <kwd>information extraction</kwd>
        <kwd>opinion mining</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Subjectivity is inherently encoded in language and involves expressions of the speaker’s position,
attitude, and feelings towards the uttered message [17]. Thus, identification of articles written from a
subjective perspective of the author involves detection of stance markers: words that express some
form of evaluation or judgement (e.g. words denoting emotional valence), pronouns or modal verbs
and passive constructions [16]. This work makes use of syntactic and semantic features that are fed
to machine learning algorithms and large language models (LLMs) for a binary detection of news
articles written from a subjective versus objective perspective of the author [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The datasets have been
provided by the organizers of the CLEF 2024 CheckThat! lab [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which is the international contest on
challenging classification and retrieval problems. The aim of this competition is to advance the field of
information retrieval from text. To this end, a new approach to subjectivity detection is proposed: a
Transformer-based model that makes use of both text content and additional meta features derived
from articles. Such an approach shows more consistent results than simple transfer learning (TL; adding
a classification layer on top of BERT encoder) fed with textual content only. The current work describes
the approach by the eevvgg team for Task 2: Subjectivity in News Articles of the CLEF 2024 CheckThat!
lab in English news articles.
      </p>
      <p>
        It is a common approach in information retrieval research to experiment with diferent text
representation models, as in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] where conversion of text with TF-IDF (Term Frequency Inverse Document
Frequency) algorithm outperformed models with so-called Count Vectorizer, that is, a simple frequency
count of particular terms in each text sample. Influence of preprocessing techniques on classification
performance of deep learning models was tested in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Machine learning methods are among
fundamental methods used in the field of natural language processing [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. They are used, for example, in web
search engines for information retrieval [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Thereby, a user looking for specific information gets results
relevant to the searched topic. Previous studies investigated performance of several machine learning
models: logistic regression, SVM, and Naive Bayes for text classification tasks such as recognition of
political afiliation of the US presidential candidates from their presidential campaign speeches [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Contribution of this work is three-fold: first, examination of machine learning vs. transfer learning
approaches for information extraction; second, impact of feature normalisation on classification
performance; third, influence of training data size on detection performance. Models for detection of
subjective opinions versus objective facts could be utilised, for example, in studies on information
verification (detection of fake news, which are a mixture of subjective opinion and facts [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]).
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Diferentiation between subjective opinions and objective facts compose a basis of proper journalism.
Subjective reporting of news might reflect bias of the author which automatic solutions can identify
and tag properly to reduce the spread of fake news, for example. Several work have experimented with
subjectivity detection methods from text. Unified method for detection of subjectivity in multilingual
text content was proposed in [12]. Fine-tuned ELECTRA large model outperformed other large language
models (such as BERT and RoBERTa) for subjectivity analysis in text in the context of fake news detection,
achieving 0.983 accuracy [20]. Such analyses have proved to be useful for detecting fake news with
a lexicon-based approach [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] as well as in opinion mining using deep learning techniques [18]. The
current work extends these studies by combining previous approaches and testing the proposed model
in several experimental settings. Team eevvgg proposes a deep learning model that combines the
ifne-tuned BERT encoder [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and lexicon-based method for extracting linguistic markers of subjectivity.
The proposed architecture is called BERTd and is tested against other BERT-based models and machine
learning algorithms in two experimental conditions: on a smaller vs. a larger set of data (n=667 vs.
n=1511 training samples, and n=166 vs. n=484 test samples, respectively); and on raw vs. normalised
values of linguistic meta features.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Material</title>
        <p>This work deals with binary detection of subjectivity from text material: whether a sentence expresses
a subjective view of the author (SUBJ) or presents an objective view on the topic (OBJ). The paper
proposes solutions for subjectivity extraction for English data. Initial training data comprises 833 text
samples (data units) from newspaper articles: short pieces of text of up to 100 words. Development of
a text classification system starts with pre-processing of the data, then splitting it into training and
testing sets, extraction of features, model training and model validation. Then, the dataset was divided
into two splits: 80% for training purposes and 20% for testing. Results of training and evaluation on this
set (called small set) are reported in Table 1. Evaluation and training is conducted also on a bigger set
oficial test set - released by the CheckThat! organizers after submission deadline and training data
comprising all available train and dev sets - 1511 in total; evaluation is conducted on 484 samples from
the oficial test. Results of training and evaluation on this set (called big set) are reported in Table 2.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Text Preprocessing</title>
        <p>Usually, the first step in information retrieval tasks is to represent the text using a certain model. A
common approach is to represent a document as a vector of features - most simple representations of
text include Bag-of-Words (BOW) models. Regarding machine learning algorithms TF-IDF method is
employed (with max_df=0.75, min_df=2). Transfer learning approaches utilise BERT encoder (BERT
base uncased) as a text representation method. In order to improve the predictive performance of these
models, several meta features were constructed from the textual content of news articles using the
concept of feature engineering [14]. It involves the application of transformation functions on given
features to generate new ones. In order to extract such features from text, it needs to be normalised. Data
cleaning involved 3 steps: conversion of text into lowercase, removal of stop-words and punctuation
symbols. Then, the text was lemmatised, that is, words were converted to their dictionary forms. Finally,
linguistic features were extracted from the clean and lemmatised text of articles. In total, 16 syntactic
features (stance markers) were extracted from text samples, specifically, frequency of occurrence of
each category of terms in a given text sample. Specific terms that belong to each category are specified
below:
1. Subject pronouns: I, you, he, she, it, we, you, they;
2. Object pronouns: me, you, him, her, it, us, you, them;
3. Possessive pronouns: mine, yours, his, hers, its, ours, yours, theirs;
4. Demonstrative pronouns: this, these, that, those
5. Interrogative pronouns: who, whom, which, what
6. Relative pronouns: who, whom, that, which, whoever, whichever, whomever;
7. Indefinite pronouns: all, another, any, anybody, anyone, anything, each, everybody, everyone,
everything, few, many, nobody, none, one, several, some, somebody, someone;
8. Reflexive pronouns: myself, yourself, himself, herself, ourselves, yourselves, themselves
9. Modal verbs: must, shall, will, should, would, can, could, may, might;
10. Obligation verbs: need, have to, must, might, may, has to, shall;
11. Frequency adverbs: hardly, ever, rarely, scarcely, seldom, never, sometimes, often, always, usually,
normally;
12. Comparison adverbs: bad, badly, worse, worst, good, better, well, best, far, farther, further, farthest,
furthest, little, less, least, few, somehow;
13. Reporting verbs: advise, agree, challenge, claim, decide, demand, encourage, invite, ofer, persuade,
promise, refuse, remind, say;
14. Pronouns: a sum of features 1-8;
15. Emotive words: words associated with an expression of emotions marked as such in the lexicon
of emotion-laden terms [13];
16. Polarising words: words associated with inducing social polarisation (dividing the society into
‘us’ versus ‘them’ groups) marked as such in the lexicon of polarising terms [19];</p>
        <sec id="sec-3-2-1">
          <title>The summary of preprocessing procedure is illustrated in Figure 1.</title>
          <p>
            3.3. Tools
In the light of available algorithms, scikit-learn map of estimators is followed in order to chose four of
them: Naive Bayes (NB), logistic regression (LR), decision trees (DT) and decision forests (DF). The Naive
Bayes is one of the simplest and most popular models in the field of supervised machine learning [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ]
and amongst the most eficient and efective classifiers [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. Classification with the use of this estimator
is based on the calculated probabilities: probability of a certain label for a given data point is estimated
through multiplying the probability of this label by a sum of probabilities of all features describing
this data point given the label [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ]. The NB algorithm learns probabilities based on prior distribution
across classes from the training data following the assumption that all features are independent. SVM
models learn to categorise data into separate classes by building a margin in the feature space that
minimizes the distance between each class and that margin [21]. The logistic regression algorithm
calculates the log-odds (converted into probabilities) of an event as a linear combination of one or more
independent variables. The Decision Tree model predicts the value of a target variable by learning
decision rules, which are inferred from the training data. DT builds a tree-like structure from these
rules by splitting data into subsets based on the values of particular features until a stopping criterion
is met. DTs have two main advantages of being simple to understand and interpret [15]. A decision tree
forest is an ensemble learning method that combines outputs of multiple decision trees to reach the
ifnal result. Finally, the suitability of large language models for subjectivity detection is investigated.
Specifically, BERT uncased model 1) that produces contextualised text embeddings and achieves state of
the art results in most information retrieval tasks. Transfer learning paradigm is utilised for training
BERT-based models. Tensorflow and transformers libraries are employed for their implementation.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.4. Experimental Settings</title>
        <p>Machine Learning. Default settings of hyper-parameters set by the authors of the scikit-learn library
are employed in machine learning estimators. TF-IDF method is utilised as a text representation method.
Textual features are combined with (normalised) meta features described in Section 2.1 and fed to these
algorithms.</p>
        <p>Transfer Learning. Regarding TL with LLMs, BERT is utilised for as a text encoder, specifically the
CLS token. Then, additional layers are added on top of embeddings returned by the BERT-encoder. Two
BERT-based models are proposed: BERTs comprises of two fully-connected layers2 of size 128 and a 0.5
dropout rate between them; in BERTd two fully-connected layers of size 246 and 32, separated by a 0.4
dropout layer are attached to the BERT encoder. In addition BERTb is utilised a baseline model with
only a classification layer added on top of the BERT encoder. Cross-entropy is employed as the loss
function and rectified linear unit (ReLU) as the activation function in all hidden layers. Classification
layer comprises of two units, which in combination with the softmax function returns probabilities
that a given text sample belong to class "OBJ" and "SUBJ". Learning rate is set to 5e-5 as advised by the
authors of the transformers library. BERT-based models are trained for either 2, 3 or 4 epochs and the
best result is reported. Tensorflow implementation of these networks and functions is employed.
Evaluation metrics. Three metrics are employed for evaluation purposes: weighted 1 (Eq. 2),
macro-averaged 1 (Eq. 3) and accuracy (Eq. 4).</p>
        <p>1 =</p>
        <p>2 ×  
2 ×   +   +  
 ℎ 1 =
  1 =
∑︀=1 ×  ×  1</p>
        <p>∑︀=1 ×  1</p>
        <p>+  
 =</p>
        <p>+   +   +  
where TP: true positives, FP: false positives, TN: true negatives, FN: false negatives.
(1)
(2)
(3)
(4)</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>Results on a small train set, reported in Table 1, indicate that BERTb, a simple transfer learning
approach, outperforms all other classifiers. Nonetheless, all tested models perform above baseline using</p>
      <sec id="sec-4-1">
        <title>1https://huggingface.co/google-bert/bert-base-uncased 2https://keras.io/api/layers/core_layers/dense/</title>
        <p>prior label distribution. Normalisation of additional features boosts performance for ML algorithms
(NB, LR, DT, RB), although BERT-based models note minor diferences in performance. All models
developed with a TL approach outperform all ML algorithms achieving, on average, 27% higher results
(23% for normalised meta features and 30% for models without normalisation or meta features) in terms
of macro 1.</p>
        <p>The proposed BERTd model (oficially submitted to the task), consisting of two hidden layers and
fed with data without normalisation of meta features, outperforms other solutions in turn on the
bigger set (see Table 2). The diference in performance between ML and TL approaches decreases to 20%
- due to lower macro 1 of two TL models: BERTb and BERTs. Compared to BERTb and BERTs, BERTd
notes smaller diferences in performance between small and big training sets. Thus, its performance is
more stable across data than BERTb and BERTs. Nonetheless, all models outperform a baseline classifier
using class prior distribution. All TL models and the LR classifier achieve also higher macro 1 than
the baseline provided by the CheckThat! organisers.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Once again, transfer learning outperformed simpler machine learning approaches for information
extraction from text. All BERT-based models (with or without syntactic features) achieved substantially
higher results for a binary detection of subjectivity than 4 ML algorithms. BERTd fed with additional
features (syntactic features of marker) shows more consistent performance than a baseline BERTb
model comprising of BERT encoder and a classification layer. Normalisation of meta features was found
to boost performance for ML models. Increase of training data size had almost no impact on prediction
performance for ML models and a negative influence for TL solutions in terms of macro 1 metric.</p>
    </sec>
    <sec id="sec-6">
      <title>Limitations</title>
      <p>Baselines. The current work would benefit from a more thorough analysis of results of the proposed
models against other systems developed on the employed dataset.</p>
      <p>Generalisability of performance. The proposed architecture notes satisfactory performance on the
utilised dataset (ranking 8th in the oficial leaderboard of the competition on English data), however, its
robustness is yet to be tested, for example, in scenarios with diferent training datasets or multilingual
data.</p>
      <p>Ablation studies. Ablation analysis in future work could measure the impact of individual features
on the final performance of the proposed subjectivity detectors.
International Conference on Information Integration and Web-based Applications &amp; Services (pp.
15-24).
[12] Karimi, S., &amp; Shakery, A. (2017). A language-model-based approach for subjectivity detection.</p>
      <p>Journal of Information Science, 43(3), 356-377.
[13] Mohammad, S. M., &amp; Turney, P. D. (2013). Crowdsourcing a word–emotion association lexicon.</p>
      <p>Computational intelligence, 29(3), 436-465.
[14] Nargesian, F., Samulowitz, H., Khurana, U., Khalil, E. B., &amp; Turaga, D. S. (2017, August). Learning</p>
      <p>Feature Engineering for Classification. In IJCAI (Vol. 17, pp. 2529-2535).
[15] Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., . . . others. (2011).</p>
      <p>Scikit-learn: Machine learning in Python. The Journal of machine learning research, 12, 2825-2830.
https://scikit-learn.org/stable/
[16] Reilly, J., Zamora, A., &amp; McGivern, R. F. (2005). Acquiring perspective in English: the development
of stance. Journal of pragmatics, 37(2), 185-208.
[17] Ruggeri, F., Antici, F., Galassi, A., Korre, K., Muti, A., &amp; Barrón-Cedeño, A. (2023). On the Definition
of Prescriptive Annotation Guidelines for Language-Agnostic Subjectivity Detection. Text2Story at
ECIR, 3370, 103-111.
[18] Sagnika, S., Mishra, B. S. P., &amp; Meher, S. K. (2021). An attention-based CNN-LSTM model for
subjectivity detection in opinion-mining. Neural Computing and Applications, 33(24), 17425-17438.
[19] Simchon, A., Brady, W. J., &amp; Van Bavel, J. J. (2022). Troll and divide: the language of online
polarization. PNAS nexus, 1(1), pgac019.
[20] Vieira, L. L., Jeronimo, C. L. M., Campelo, C. E., &amp; Marinho, L. B. (2020, November). Analysis of the
subjectivity level in fake news fragments. In Proceedings of the Brazilian Symposium on Multimedia
and the Web (pp. 233-240).
[21] Vinodhini, G., &amp; Chandrasekaran, R. M. (2012). Sentiment analysis and opinion mining: a survey.</p>
      <p>International Journal, 2(6), 282-292.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Acharya</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crawford</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Maduabum</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>A nation divided: Classifying presidential speeches</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Aggarwal</surname>
            ,
            <given-names>C. C.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Zhai</surname>
            ,
            <given-names>C</given-names>
          </string-name>
          . (Eds.). (
          <year>2012</year>
          ).
          <article-title>Mining text data</article-title>
          . Springer Science &amp; Business Media.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Alshdaifat</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alshdaifat</surname>
            ,
            <given-names>D. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alsarhan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hussein</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>El-Salhi</surname>
            ,
            <given-names>S. M. D. F. S.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>The efect of preprocessing techniques, applied to numeric features, on classification algorithms' performance</article-title>
          .
          <source>Data</source>
          ,
          <volume>6</volume>
          (
          <issue>2</issue>
          ),
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Antici</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruggeri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galassi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Korre</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bardi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Barrón-Cedeño</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2024</year>
          , May).
          <article-title>A Corpus for Sentence-Level Subjectivity Detection on English News Articles</article-title>
          .
          <source>In Proceedings of the 2024 Joint International Conference on Computational Linguistics</source>
          ,
          <article-title>Language Resources and Evaluation (LREC-COLING</article-title>
          <year>2024</year>
          )
          <article-title>(pp</article-title>
          .
          <fpage>273</fpage>
          -
          <lpage>285</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Barrón-Cedeño</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          et al. (
          <year>2024</year>
          ). The CLEF-2024 CheckThat! Lab:
          <string-name>
            <surname>Check-Worthiness</surname>
            , Subjectivity, Persuasion, Roles, Authorities, and
            <given-names>Adversarial</given-names>
          </string-name>
          <string-name>
            <surname>Robustness</surname>
            . In: Goharian,
            <given-names>N.</given-names>
          </string-name>
          , et al.
          <source>Advances in Information Retrieval. ECIR 2024. Lecture Notes in Computer Science</source>
          , vol
          <volume>14612</volume>
          . Springer, Cham. https://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -56069-9_
          <fpage>62</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Basarkar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Document Classification using Machine Learning</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Bird</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Loper</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>Natural language processing with Python: analyzing text with the natural language toolkit.</article-title>
          <string-name>
            <surname>O'Reilly Media</surname>
          </string-name>
          , Inc.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>De</given-names>
            <surname>Grandis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Pasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            , &amp;
            <surname>Viviani</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          (
          <year>2019</year>
          ,
          <article-title>July). Multi-criteria decision making and supervised learning for fake news detection in microblogging</article-title>
          .
          <source>In Workshop on Reducing Online Misinformation Exposure</source>
          (pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M. W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Grimmer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Stewart</surname>
            ,
            <given-names>B. M.</given-names>
          </string-name>
          (
          <year>2013</year>
          ).
          <article-title>Text as data: The promise and pitfalls of automatic content analysis methods for political texts</article-title>
          .
          <source>Political analysis</source>
          ,
          <volume>21</volume>
          (
          <issue>3</issue>
          ),
          <fpage>267</fpage>
          -
          <lpage>297</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Jeronimo</surname>
            ,
            <given-names>C. L. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marinho</surname>
            ,
            <given-names>L. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Campelo</surname>
            ,
            <given-names>C. E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Veloso</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp; da Costa Melo,
          <string-name>
            <surname>A. S.</surname>
          </string-name>
          (
          <year>2019</year>
          , December).
          <article-title>Fake news classification based on subjective language</article-title>
          .
          <source>In Proceedings of the 21st</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>