<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Exploring Data Augmentation for Classification of Climate Change Denial: Preliminary Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jakub Piskorsk</string-name>
          <email>jpiskorski@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>NikolaosNikolaidi</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>NicolasStefanovitc</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>BonkaKotseva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>IreneVianin</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sopho Kharaz</string-name>
          <email>sopho.kharazi@ext.ec.europa.e</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jens P. Linge</string-name>
          <email>jens.linge@ec.europa.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop Proceedings</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CRI</institution>
          ,
          <addr-line>Luxembourg</addr-line>
          ,
          <country country="LU">Luxembourg</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>European Commission Joint Research Centre</institution>
          ,
          <addr-line>Ispra</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>In: R. Campos, A. Jorge, A. Jatowt, S. Bhatia, M. Litvak (eds.): Proceedings of the Text2Story'22 Workshop</institution>
          ,
          <addr-line>Stavanger</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Piksel SRL</institution>
          ,
          <addr-line>Ispra</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Polish Academy of Sciences</institution>
          ,
          <addr-line>Warsaw</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Trasys International</institution>
          ,
          <addr-line>Brussels</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <abstract>
        <p>In order to address the growing need of monitoring climate-change denial narratives in online sources, NLP-based methods have the potential to automate this process. Here, we report on preliminary experiments of exploiting Data Augmentation techniques for improving climate change denial classification. We focus on a selection of both known techniques, and augmentation transformations not reported elsewhere that replace certain type of named entities with high probability of preserving labels. We also introduce a new benchmark dataset consisting of text snippets extracted from online news labeled with ifne-grained climate change denial types.</p>
      </abstract>
      <kwd-group>
        <kwd>text classification</kwd>
        <kwd>climate change denial</kwd>
        <kwd>machine learning</kwd>
        <kwd>data augmentation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>(J. P. Linge)
(specific name-entity type replacements) with a focus on transformations with high probability
of preserving labels. The main drive behind this research is two-fold: First, it emerges from
the need to rapidly develop a production-level component for CC denial text classification for
Europe Media Monitor (EM M1,)a large-scale media monitoring platform used by EU institutions,
and, secondly, from the scarcity of annotated data for the task at hand. The experiments reported
in this paper build mainly on top of the only publicly available text corpus of CC contrarian
claims, which is labeled using a fine-grained taxonomy presented1i]n. W[e also present a
preliminary evaluation of a some models on a new EMM-derived news snippet corpus reusing
the same taxonomy. The findings contained in this paper are not of general nature, but rather
specific to the exploited data and the domain, paving the way for future in-depth explorations.</p>
      <p>The paper starts with an overview of related work in Se2.ctNieoxnt, the DA techniques
exploited in our study is described in Sect3i,ownhereas Section4 introduces a news-derived corpus
of text snippets related to CC denial. The DA techniques performance evaluation is presented
in Section5. Section6 provides detailed analysis of the behaviour of two specific named-entity
replacement-based DA techniques. Finally, we present our conclusions in S7e.ction</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Only recently the CC debate has received more attention in the NLP community in the context
of developing solutions for making sense of the vast amount of textual data produced on this
topic [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. A corpus of manually-tagged blog posts on CC in terms of scepticism and acceptance
of CC is presented in3[]. In 2016 a SemEval task on stance detection of tweets, where ”CC is a
real concern” was organize4d].[In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] an annotated news corpus for stance toward ”climate
change is a real concern” and related experiments are presented, wh6]erineatsro[duced a
dataset for sentence-based climate change topic detection. Fi7n]arlelpyo, r[ted on a collection
of tweets used to study the public discourse around CC.
      </p>
      <p>
        To the best of our knowledge only two textual corpora with CC denial and disinformation
labels exists, namely, the corpus of ca. 30K text paragraphs containing contrarian claims about
climate change extracted from conservative think-tank websites and contrarian blogs (4C
corpus) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], and a collection of ca. 500 news articles with known CC misinformation scraped
from web pages of CC counter movement organisatio8n].s G[iven that the latter corpus is
not publicly accessible at the moment, we exploit the former 4C corpus, and the associated
taxonomy of climate contrarianism in our study.
      </p>
      <p>
        Data Augmentation (DA) is a family of techniques aiming at creation of additional training
data in order to alleviate problems related to insuficient and imbalanced data and low data
variability, with the overall goal of model performance improvement. Recently, DA has gained
attention in the NLP domain, and a wide range of DA techniques has been elaborated and
explored, including, i.a., simple word substitution, deletion, and inse9]r,tsiuobn-s[tructure
substitution10[], back-translatio1n1[
        <xref ref-type="bibr" rid="ref12">, 12</xref>
        ], contextual augmentatio1n3][, data noising1[
        <xref ref-type="bibr" rid="ref15 ref4">4, 15</xref>
        ],
injection of noise into the embedding spa1c6e][, interpolating the vector representations of
text and label1s7[], etc. A survey on DA techniques for text classification is presente1d8]i,n [
whereas [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] provides a more general overview of DA in the broader area of NLP .
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Data Augmentation</title>
      <p>For the sake of carrying out DA experiments we have selected a range of known and 2 variants
of some known techniques, in particular, focusing on transformations with high probability of
preserving labels by the automatically created instances. The list of DA techniques encompasses:
COPY: simply creates copies of the existing instances in the training dataset.
DATE: randomly changes all dates, e.g., month and day-of-the-week names.</p>
      <p>DEL-ADJ-ADV: deletes up to a maximum of 1/3 of all adjectives and adverbs in the text,
provided that they are preceded by nouns and verbs resp. Here, the assumption is that such
transformation preserves the label assigned to the text.</p>
      <p>
        PUNCT: inserts various punctuation marks randomly selected from (‘.’, ‘;’, ‘?’, ‘:’, ‘!’, ‘,’) into
randomly selected positions in the text, where the number of insertions is a randomly selected
number between 1 and 1/3 of the length of the text (in words). This simple DA technique
introduced recently i1n5[] proved to outperform many other simple DA techniques.
GEO: randomly replaces all occurrences of toponyms referring to a populated place with another
randomly chosen toponym fromGeonames-based2 gazetteer of about 200K populated cities.
PER-ORG: randomly replaces occurrences of mentions of person and organisation names
matched using theJRC Name Variant database2[0] (containing large fraction of entities whose
mentions appear in the news) with some other names therefrom (not spelling variants of the
replaced names). The current versionJRoCf Name Variant contains circa 3 million names.
SYN: randomly replaces verbs and adjectives with their synonyms. It picks the top-10 tokens
(verbs/adjectives) whose deletion maximizes the cosine distance from the resulting sentence’s
embedding to that of the original sentence and replaces them with semantically close words. For
the first part, we exploit USE embeddin2g1[] and for the second, we approximate the semantic
proximity of words with wikipedia pre-traiFnaesdtText embeddings [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]3.
      </p>
      <p>SYN-REV: same process as above, but difers in picking the top-10 tokens whose deletion
minimizes the cosine distance of the sentence’s embedding.</p>
      <p>BACK-TRANSL: consists of translating the input text to some other language and then
translating back the translation into Eng11li,s1h2][. Here, we translated to French, German and
Polish and then back to English using an in-house NMT-based solu2t4i]o.n [
Some examples of the application of the DA techniques enumerated above are provided in
Table8 in AnnexA. While most of these techniques were reported elsewhere, GEO and
PERORG, i.e., replacement of specific types of named entities, to the best of our knowledge, were not
explicitly explored. Based on empirical observations, the application of these transformations
result in label preservation with high probability, although the transformed texts might appear
‘unrealistic’ due to random name replacement. Furthermore, since the replacement is based on
a lexicon look-up, the transformation might result in replacing entities of other type by mistake,
but, again, based on empirical observations, this does not have high impact on the label.</p>
      <p>Additionally, we explored ways of combining the DA techniques enumerated above, incl.: (a)</p>
      <sec id="sec-3-1">
        <title>2https://www.geonames.org/ 3We exploit thGeensim interface2[3]</title>
        <p>ALL: combination the results of all the above DA techniques created separaAtLelLy-,K(bB):,
a variant of ALL, but combining only DA techniques based on knowledge-based resources,
i.e., PUNCT, DEL-ADJ-ADV, DATE, GEO and PER-ORG, (c)ALL-KB-STACKED: resulting
from running the techniques used in ALL-KB in a pipeline (in the order as above) that modifies
progressively the same input text, andB(EdS)T-3 combination of the 3 DA techniques, whose
results were merged (not stacked), and which yield best gain in performance (see S5e).ction</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. EMM-derived CC denial text snippet corpus</title>
      <p>
        In order to establish a benchmark corpus for the news domain and to test the classification
performance, we relied on EMM. Articles taken from a limited set of news sources that
disinformation experts had identified as frequently spreading misinformation. In order to limit the
dataset to articles on CC, we queried for articles containing keywords related to the topic such
as: ‘climate change’, ‘global warming’, ‘greenhouse gas[es]’, ‘greenhouse efect[s] ’ and limited the
publication date to the whole of 2021. Out of these, a random subset of 2500 articles was sampled.
For each article, we generated a snippet made of the title and of up to the first 500 characters.
The corpus was manually annotated by five disinformation experts, using the Codebook defined
in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. 1118 snippets were annotated, 42.7% of which are tagged with a class indicating a CC
denial narrative, while the second half has been taggNeodcalasim, i.e, not containing any
CC denial claim captured by the Codebook. In some snippets, while inflammatory language
superficially similar to CC denial was used, the texts actually embrace polemical stance on
CC inaction. When stance was ambiguous, the snippet was discarded, whereas the remaining
snippets containing activists stance were assigned theNloacblaeilm.
      </p>
      <p>The statistics of the current version of the c4oarpreusprovided in AnneAx in Table7.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Classification Experiments</title>
      <p>
        We have experimented with two ML paradigms, namely: (a) linear SVM using the algorithm
described in 2[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] andLiblinear librar5y, with 3-6 character n-grams as binary features, using
vector normalization a n=d1.0 resulting from parameter optimization, and (b) RoBERTA
architecture26[] using batch size=32, learning rate 1e-5 andclass weighting.
      </p>
      <p>Prior to carrying out ML experiments we cleaned the original 4C c1o]rdpuues t[o some
problems, i.a., (a) some entries were included in both training and test data, and often having
diferent labels, and (b) some entries were corrupt, i.e., missing texts, non parseable content.
We used this modified version of the 4C corpus, containing ca. 30 entries less. The 4C dataset
is highly imbalanced, i.e., more than 60% of the instances are labeNleodcalasim, whereas 14
classes constitute ca. 1-2% of the entire dataset each (see6TinabAlnenexA for statistics.).</p>
      <p>The results of the evaluation of SVM and RoB E RTaon the 4C corpus without any DA
are presented in Tab1l,ewhere we explored SVM both with and without class weighting. The
performance of the baseline RoBERTa is similar to the one of its counterpart report1e].d in [</p>
      <p>As regards DA techniques, we have augmented all instances CofCa-ldlenial classes, whereas
theNo claim class was not augmented. Each to-be-augmented instance was augm en∈ted
{1, 2, 4} times and the experiments have been repeated 3 times. The gain/loss obtained for
SVM4Please note that this corpus is ongoing active development and will be continuously extended.
5https://www.csie.ntu.edu.tw/~cjlin/liblinear
and RoBERTa-based models for all DA techniques is reported in T2aabnlde3 resp., with the
best results per measure and number of augmentations marked in bold. In all experiments all
original training data was used as well. BEST-3 refers to a combination of 3 DA techniques, each
run separately, which yield best gain in performance and were: (a) PUNCT, BACK-TRANSL,
GEO for SVM, and (b) PUNCT, GEO, PER-ORG for weighted SVM and RoBERTa.</p>
      <p>1 augmentation
Accuracy macro  1
gain gain</p>
      <p>2 augmentations
Accuracy macro  1
gain gain</p>
      <p>4 augmentations
Accuracy macro  1</p>
      <p>gain gain
+0.6 (+2.1)
-0.4 (+0.9)
-0.1 (+1.1)
+0.4 (+1.9)
+0.1 (+2.0)
+0.6 (+1.8)
-0.3 (+0.7)
-1.5 (-1.7)
-0.8 (+2.0)
+1.6 (+2.3)
+0.4 (+2.5)
-0.1 (+2.7)
+1.4 (+3.8)
+3.0 (+0.6)
+0.3 (+0.1)
+2.1 (+0.2)
+3.4 (+0.2)
+2.0 (+0.6)
+1.4 (+0.4)
+3.0 (+0.3)
+2.4 (-0.1)
+1.7 (+0.6)
+4.0 (+1.0)
+2.9 (+0.8)
+4.3 (+0.8)
+3.7 (+1.1)
+0.8 (+1.8)
-0.2 (+0.5)
+0.3 (+1.2)
+1.1 (+1.0)
+0.5 (+2.1)
+0.7 (+1.5)
-0.2 (+0.4)
-0.9 (-0.2)
-0.3 (+2.1)
+1.0 (+2.8)
+0.9 (+2.5)
+0.5 (+2.7)
+1.4 (+3.8)
+3.9 (+0.3)
+0.5 (0.0)
+3.1 (+0,4)
+4.3 (+0.5)
+3.0 (+0.9)
+2.1 (+0.3)
+3.9 (+0.4)
+3.9 (0.0)
+2.2 (+0.7)
+4.6 (+0.6)
+3.8 (+0.8)
+4.6 (+0.7)
+4.6 (+0,7)
+0.9 (+1.7)
-0.3 (+0,4)
+0.3 (+1.1)
+1.2 (+1.6)
+0.6 (+2.7)
+0.6 (+1.1)
-0.6 (+0.1)
-0.5 (-0.3)
-0.1 (+2.4)
0.1 (+2.1)
+0.7 (+2.6)
+0.2 (+1.9)
+0.9 (+3.1)</p>
      <p>As regards weighted SVM, one can observe that overall highest gain in m1awcarsoobtained
with the ALL-KB setting (+1.6) with a 1-per-instance augmentation, and BEST-3 obtained highest
gain for 2 and 4 augmentations (1.4 and 0.9 resp.). PUNCT appears to be the best stand-alone
DA technique with some gains above 1.0. Applying simple copying (COPY) beats many other
DA techniques (macr o 1 improved by up to +0.9), although it is outperformed by the ones
mentioned earlier. The two new DA techniques, i.e., GEO and PER-ORG yield positive gain
in all set-ups, while the usage of DATE, SYN, SYN-REV and BACK-TRANSL in a stand-alone
mode does not appear to be beneficial, i.e., close to zero gain or deterioration. The DA gains for
unweighted SVM are higher, but since the best setting (BEST-3) for unweighted SVM case is
worse than the weighted SVM baseline, we do not analyze it any further.</p>
      <p>As regards the RoBER Ta -based models, one can observe that DA consistently deteriorates
the accuracy on average, whereas for the most of the basic DA techniques there is little or
1 augmentation
Accuracy macro  1
gain gain</p>
      <p>2 augmentations
Accuracy macro  1
gain gain</p>
      <p>4 augmentations
Accuracy macro  1</p>
      <p>gain gain
no again at all in terms of macr1owith BACK-TRANSL exhibiting the highest gain (+0.9),
followed by PUNCT (+0.6). The composite DA techniques perform on average better, with
highest gain of 0.7 for ALL, which is higher than when applying simple COPY (0.4). Such results
are consistent with recent literature exploring data augmentation techniques with RoBERTa in
the related field of propaganda techniques classificati2o7n].[</p>
      <p>RoBERTa’s deterioration could be possibly explained by potential overfitting to the full
sentence structure due to too similar sentences, given neural networks tendency t1o8]o.verfit [
However, we also observe that this phenomenon diminishes with more augmentations. While
DATE should have the least impact on the label, it showed the most important and consistent
drop in performances. A better understanding of this behaviour requires further investigation.</p>
      <p>Interestingly, we have observed that PUNCT, SYN, and SYN-REV were the three basic DA
techniques with highest variance (up to ca. 1.0 diference in the gain for m a1carcoross diferent
experiments) and the same could be observed for the composite methods that do include these
basic DA methods. In particular, given that the simple PUNCT method performs overall best
across the diferent settings one could explore in future potential improvements that could be
gained through some tuning, e.g., limiting the positions in which punctuation signs are inserted
and/or studying whose punctuation sign insertion results in higher gains in performance.</p>
      <p>We have applied the baseline and some DA-boosted models on the EMM-derived corpus
described in Section4, whose performance is summarized in Tabl4e. The deterioration in
performance vis-a-vis 4C corpus evaluation could be mainly due to the diferent nature of the
EMM corpus (text structure and writing style). Noteworthy, the evaluation on the EMM dataset
revealed that the models trained using DA consistently outperform the baseline models. As
regards the RoBERTa-based data-augmented models the gain ranges from -0.7 to +2.8 and -0.7
to +4.6 in accuracy and macr o1 scores, respectively, with the vast majority being positive.
The boost is the result of higher recall in the DA trained models. For the sake of completeness,
the confusion matrix for the RoBE R Ta model boosted with BACK-TRANSL augmentation
(reported in Tabl4e) is provided in Figure1 AnnexA</p>
    </sec>
    <sec id="sec-6">
      <title>6. Data Augmentation Impact on Reducing the Bias</title>
      <p>In order to better understand the behavior of the DA techniques relying on proper name
replacement, namely GEO and PER-ORG, we performed additional experiments with alternate
versions, and analysing the distribution of names entities. This is motivated by the finding
that texts containing disinformation are often very specific about the entities involved. These
alternate techniques are characterized by a diferent sampling strategy of the entities to be
inserted. In contrast to the GEO and PER-ORG experiments, the replaced named entities are
not taken from a larger pool of entities, but instead, are taken from the pool of the entities
that are detected in the texts. We respectively define the additional expeGriEmOe-nStPs and
PER-ORG-SP which correspond to the GEO and PER-ORG experiments using this modified
sampling on the CC-denial classes onGlyE;O-SP-ALL and PER-ORG-SP-ALL, where this
randomization procedure is applied to the CC-denial classes as well aNs otoclatihme class;
and finally GEO-SP-STRICT andPER-ORG-STRICT, where the instances of all classes are
perturbed and only perturbed data is used. These experiments were only performed with
weighted SVM, using only one augmentation. We report the results in5.TWabelaelso compare
the augmented dataset and the original dataset using the Jensen-Shannon (JS) divergence on
two distributions: (a) of the replaced entities, and (b) the labels associated with these entities.</p>
      <p>In the GEO and PER-ORG experiments, the entities in the instances of the CC-denial classes
were replaced with entities drawn from a much larger pool, practically removing these original
entities from the augmented data. The clearly lower performance of the *-STRICT experiment,
notably in terms of macro1 seems to indicate that some classes rely heavily on the presence of
certain entities in order to be correctly predicted. This experiment is the only one not containing
the original data at all, and the distribution of replaced entities diverges the most from the
original dataset. Most of the errors are due to CC-denial texts being preNdoicctlaeidma, swith
the classes 4_* having the most issues, this is coherent as these classes are the most linked to
policies, and therefore to the corresponding actors.</p>
      <p>The *-SP experiments, where only CC-denial classes get augmented, show a small increase
in performances. The increase in performance is notable in the *-SP-ALL experiments, where
theNo claim class also gets augmented. The distribution of entities diverges more than in
the case of *-SP, but the distribution of labels associated with these entities diverges less. The
combination of both the original dataset and the fully transformed one seems to yield the
best compromise between generalization and fitting to particular entities in the test dataset.
Exploring this interplay is an interesting direction for future works. Randomly swapping named
entities could change an actual disinformation claim into factual information or vice versa.
It is out of the scope of the classifier to deal with fact checking, however, it is important to
reckon the competing interest between a classifier that generalises well to unseen claims on
new entities and better fitting to the known narratives.</p>
      <p>The *-SP experiments exhibit a performance on par or lower than the their equivalents
without their characteristic sampling. For GEO-SP there is a clear performance gap with respect
to GEO in terms of macro1. The reason why the divergence of GEO appears lower than
GEO-SP is because it does not take into account newly introduced entities in GEO. Overall,
both GEO, which introduces new entities, and GEO-SP, which changes the distribution of labels
associated with existing entities, tend to improve the  m1aacnrod accuracy.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>We reported on preliminary experiments of using DA techniques for improving climate change
denial classification. The evaluation on the 4C corpus yielded a boost with data augmentation
up to 1.6 and 0.9 gain in macr o1 for SVM- and RoBERTa-based classifiers resp. For the vast
majority of the DA techniques respective SVM-based models resulted in gain, whereas for
most of the RoBERTa-based models a loss was observed. Analysing the new EMM-derived test
dataset introduced in this paper with ca. 1K snippets, DA techniques lead to up to 4.6 point
gains in macro 1 vis-a-vis baseline model. The overall performance is nevertheless worse than
on the 4C corpus, which was expected due to the diferent nature of the sources considered.</p>
      <p>We provided a more in-depth analysis of the behaviour of two DA techniques not reported
earlier, which randomly replace toponyms and person/organisation names, and which were
among the ones that resulted in higher gains in m a1crfoor SVM-based models.</p>
      <p>We believe the reported findings will boost the NLP research in the climate change domain. We
also make the cleaned version of the 4C and the new EMM-derived corpus publicly acc6e.ssible</p>
      <sec id="sec-7-1">
        <title>6https://github.com/jpiskorski/CC-denial-resources</title>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>A. Supplementary Information</title>
      <p>The statistics for the 4C (Contrarian Claims about Climate Change) and the news-derived text
snippet corpus are presented in Ta6blaend 7 resp. Please note that both datasets cover only a
fraction of types (18 out of 27) of the CC contrarian claim taxo1n]o.my [</p>
      <p>In Istanbul, the snow could easily reach up to 30 cm in April, Mayor Kadir Topba announced.</p>
      <p>In Istanbul, the snow could reach up to 30 cm in June, Mayor Kadir Topba announced.</p>
      <p>In Istanbul, the snow; could easily reach up to? 30 cm in June, Mayor Kadir Topba: announced.</p>
      <p>In Porto Alegre, the snow could easily reach up to 30 cm in June, Mayor Kadir Topba announced.</p>
      <p>In Istanbul, the snow could easily reach up to 30 cm in June, Mayor Stephen King announced.</p>
      <p>In Istanbul, the snow could easily be up to 30 cm in June, Mayor Kadir Topba said.</p>
      <p>In Istanbul, Mayor Kadir Topba announced that the snow could easily be up to 30 cm high in June.
model trained using augmented data with</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Coan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Boussalis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cook</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Nanko</surname>
          </string-name>
          ,
          <article-title>Computer-assisted classification of contrarian claims about climate change</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>11</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Stede</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Patz</surname>
          </string-name>
          ,
          <article-title>The climate change debate and natural language processing</article-title>
          ,
          <source>in: Proceedings of the 1st Workshop on NLP for Positive Impact</source>
          , Association for Computational Linguistics, Online,
          <year>2021</year>
          , pp.
          <fpage>8</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Diakopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Elgesem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salway</surname>
          </string-name>
          ,
          <article-title>Identifying and analyzing moral evaluation frames in climate change blog discourse</article-title>
          .,
          <source>in: Proceedings of the Eighth International AAAI Conference on Weblogs and Social Media</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>583</fpage>
          --
          <lpage>586</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mohammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kiritchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sobhani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          , C. Cherry, SemEval
          <article-title>-2016 task 6: Detecting stance in tweets</article-title>
          ,
          <source>in: Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , San Diego, California,
          <year>2016</year>
          , pp.
          <fpage>31</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luo</surname>
          </string-name>
          , ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Card</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jurafsky</surname>
          </string-name>
          ,
          <article-title>Detecting stance in media on global warming</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2020</year>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F. S.</given-names>
            <surname>Varini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Boyd-Graber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ciaramita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Leippold</surname>
          </string-name>
          ,
          <article-title>Climatext: A dataset for climate change topic detection</article-title>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Rawi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>OʼKeefe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-J.</given-names>
            <surname>Bizimana</surname>
          </string-name>
          ,
          <article-title>Twitter's fake news discourses around climate change and global warming, Frontiers in Communication 6 (</article-title>
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhatia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Lau</surname>
          </string-name>
          , T. Baldwin,
          <article-title>You are right. I am ALARMED - but by climate change counter movement</article-title>
          ,
          <source>CoRR</source>
          (
          <year>2020</year>
          )
          <article-title>a</article-title>
          .rXiv:
          <year>2004</year>
          .14907.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zou</surname>
          </string-name>
          , EDA:
          <article-title>Easy data augmentation techniques for boosting performance on text classification tasks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>6382</fpage>
          -
          <lpage>6388</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Livescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gimpel</surname>
          </string-name>
          , Substructure substitution:
          <article-title>Structured data augmentation for NLP, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>3494</fpage>
          -
          <lpage>3508</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sennrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Haddow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Birch</surname>
          </string-name>
          ,
          <article-title>Improving neural machine translation models with monolingual data, in: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Berlin, Germany,
          <year>2016</year>
          , pp.
          <fpage>86</fpage>
          -
          <lpage>96</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A. W.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dohan</surname>
          </string-name>
          , M.-T. Luong,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Norouzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <article-title>Qanet: Combining local convolution with global self-attention for reading comprehension</article-title>
          ., CoRR abs/
          <year>1804</year>
          .09541 (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kobayashi</surname>
          </string-name>
          ,
          <article-title>Contextual augmentation: Data augmentation by words with paradigmatic relations, in: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          , Volume
          <volume>2</volume>
          (
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , New Orleans, Louisiana,
          <year>2018</year>
          , pp.
          <fpage>452</fpage>
          -
          <lpage>457</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. I.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lévy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jurafsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <article-title>Data noising as smoothing in neural network language models</article-title>
          ,
          <source>in: 5th International Conference on Learning Representations, ICLR</source>
          <year>2017</year>
          , Toulon, France,
          <source>April 24-26</source>
          ,
          <year>2017</year>
          , Conference Track Proceedings, OpenReview.net,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Prati</surname>
          </string-name>
          ,
          <string-name>
            <surname>AEDA:</surname>
          </string-name>
          <article-title>An easier data augmentation technique for text classification</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2021</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Punta Cana, Dominican Republic,
          <year>2021</year>
          , pp.
          <fpage>2748</fpage>
          -
          <lpage>2754</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Prati</surname>
          </string-name>
          ,
          <article-title>Adversarial training for aspect-based sentiment analysis with bert</article-title>
          ,
          <source>in: 2020 25th International Conference on Pattern Recognition (ICPR)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>8797</fpage>
          -
          <lpage>8803</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cissé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. N.</given-names>
            <surname>Dauphin</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <article-title>Lopez-Paz, mixup: Beyond empirical risk minimization</article-title>
          ,
          <source>CoRR abs/1710</source>
          .09412 (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bayer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kaufhold</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Reuter</surname>
          </string-name>
          ,
          <article-title>A survey on data augmentation for text classification</article-title>
          ,
          <source>CoRR abs/2107</source>
          .03158 (
          <year>2021</year>
          ). URL:https://arxiv.org/abs/2107.031.5a8rXiv:
          <fpage>2107</fpage>
          .
          <fpage>03158</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S. Y.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Gangal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chandar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vosoughi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mitamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Hovy</surname>
          </string-name>
          ,
          <article-title>A survey of data augmentation approaches for NLP, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>968</fpage>
          -
          <lpage>988</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          , G. Jacquet,
          <string-name>
            <given-names>R.</given-names>
            <surname>Steinberger</surname>
          </string-name>
          ,
          <article-title>Jrc-names: Multilingual entity name variants and titles as linked data</article-title>
          ,
          <source>Semantic Web</source>
          <volume>8</volume>
          (
          <year>2017</year>
          )
          <fpage>283</fpage>
          -
          <lpage>295</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>D.</given-names>
            <surname>Cer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          , S.-y. Kong,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Limtiaco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>St. John</surname>
          </string-name>
          , N. Constant, M. GuajardoCespedes, S. Yuan,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Strope</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kurzweil</surname>
          </string-name>
          ,
          <article-title>Universal sentence encoder for English</article-title>
          ,
          <source>in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</source>
          , Brussels, Belgium,
          <year>2018</year>
          , pp.
          <fpage>169</fpage>
          -
          <lpage>174</lpage>
          . URL: https://aclanthology.org/D18-2.0d2o9i:
          <fpage>10</fpage>
          .18653/v1/
          <fpage>D18</fpage>
          -2029.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Puhrsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          ,
          <article-title>Advances in pre-training distributed word representations</article-title>
          ,
          <source>in: Proceedings of the International Conference on Language Resources and Evaluation (LREC</source>
          <year>2018</year>
          ),
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>R.</given-names>
            <surname>Řehůřek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sojka</surname>
          </string-name>
          ,
          <article-title>Software Framework for Topic Modelling with Large Corpora</article-title>
          ,
          <source>in: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          ,
          <string-name>
            <surname>ELRA</surname>
          </string-name>
          , Valletta, Malta,
          <year>2010</year>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>h5t0t</lpage>
          .p://is.muni.cz/publication/884893/.en
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>C.</given-names>
            <surname>Oravecz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kolovratník</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jellinghaus</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <article-title>Eisele, etranslation's submissions to the WMT 2021 news translation task</article-title>
          , in: L.
          <string-name>
            <surname>Barrault</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Bojar</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Bougares</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Chatterjee</surname>
            ,
            <given-names>M. R.</given-names>
          </string-name>
          <string-name>
            <surname>Costa-jussà</surname>
            , C. Federmann,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Fishel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Fraser</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Freitag</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Graham</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Grundkiewicz</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Guzman</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Haddow</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Huck</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Jimeno-Yepes</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Koehn</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Kocmi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Morishita</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          Monz (Eds.),
          <source>Proceedings of the Sixth Conference on Machine Translation, WMT@EMNLP</source>
          <year>2021</year>
          ,
          <string-name>
            <given-names>Online</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <source>November 10-11</source>
          ,
          <year>2021</year>
          , Association for Computational Linguistics,
          <year>2021</year>
          , pp.
          <fpage>172</fpage>
          -
          <lpage>179</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>K.</given-names>
            <surname>Crammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          ,
          <article-title>On the learnability and design of output codes for multiclass problems</article-title>
          ,
          <source>in: Proceedings of the Thirteenth Annual Conference on Computational Learning Theory, COLT</source>
          '
          <fpage>00</fpage>
          , Morgan Kaufmann Publishers Inc., San Francisco, CA, USA,
          <year>2000</year>
          , p.
          <fpage>35</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized BERT pretraining approach</article-title>
          , CoRR abs/
          <year>1907</year>
          .11692 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1907</year>
          .116.9a2rXiv:
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>V.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sharma</surname>
          </string-name>
          , Nlpiitr at semeval
          <article-title>-2021 task 6: Roberta model with data augmentation for persuasion techniques detection</article-title>
          ,
          <source>in: Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1061</fpage>
          -
          <lpage>1067</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>