<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cicognini at ACTI: Analysis of techniques for conspiracies individuation in Italian</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giacomo Cignoni</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alessandro Bucci</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Pisa</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This report illustrates methods and results for solving SubtaskA (conspiracy detection) and SubtaskB (conspiracy topic classification) of EVALITA 2023 ACTI challenge. We employed diferent transformer-based models and an original method based on tf-idf. Results shows top performance scores over 80% for both subtasks. a radicalization process after the application of content moderation [8]. Therefore, the need for automatic modWe decided to cover the EVALITA 2023 challenge "Au- els that can detect the difusion of troublesome (or more tomatic Conspiracy Theory Identification" or ACTI for specifically) conspiratorial content has become crucial. short [2]. This challenge is about classifying whenever Transformer based models have revolutionized modern an Italian message is conspiratorial or not and, if positive, natural language processing [9, 10, 11, 12]. Indeed, they what type of conspiracy is about. Therefore the challenge are the current state of the art models in most NLP tasks is subdivided is 2 subtasks: spanning diferent fields from politics [ 13, 14], conflict prediction [15], and, of course, hate speech detection • CmoondeslpmiruasttorreiaclogCnoizneteifnat tCelleagsrsaimficaptioosnt is: ctohne- [16, 17, 18].[19]. In particular finetuning of BERT[ 20] spiratorial or not. based models for classification tasks such as sentiment analysis or topic detection has been widely studied and • Conspiracy Category Classification : the its efectiveness proved with multiple benchmarks [ 21]. model must discriminate to which conspiracy The usage of machine learning techniques for detecting theory a post belongs from a list of 4 possible conspiracy theories has been studied mainly in regard conspiracy topics: to social media texts extracted in the English language, although also classification on diferent topic of the conspiracies has been considered [22, 23].</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Conspiracy Theory</kwd>
        <kwd>Content Moderation</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Computational Social Science</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>1. Covid-Conspiracy
2. Qanon-Conspiracy
3. Flat Earth-Conspiracy
4. Pro-Russia Conspiracy</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>
        Conspiratorial content has been raising on the internet
over the past years such that some has define it as a
"Golden Age of Conspiracy" [
        <xref ref-type="bibr" rid="ref6">3</xref>
        ]. Indeed mainstream
platforms have tried to moderate the difusion of online
communities with the implementation of content moderation
known as deplatforming. However, there have been a
lot of discussion regarding the eficacy of such
interventions [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">4, 5, 6</xref>
        ].. Indeed, some identified the presence of
spillover of toxic behaviour [7] and the the presence of
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Datasets</title>
      <p>The 2 provided datasets are a collection of labeled Italian
Telegram’s messages. Both datasets were relatively clean
in regard to the text, so heavy preprocessing was not
needed.</p>
      <sec id="sec-3-1">
        <title>3.1. Subtask A dataset</title>
        <p>More specifically for Subtask A, the training dataset is a
.csv file containing:
• id: unique post identifier.
• comment_text: the text of the telegram’s
message.
• conspiratorial: a binary label that indicates if
the message is conspiratorial or not.</p>
        <p>The training dataset is composed by 1842 samples, of
which 925 with a positive conspiratorial label and 917
with a negative conspiratorial label. The hidden test set
is composed by 460 samples instead.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Subtask B dataset</title>
        <p>And for Subtask B, the training dataset is a .csv file
containing:
• id: unique post identifier.
• comment_text: the text of the telegram’s
message.
• conspiracy: a label going from 0 to 3 indicating
which conspiracy topic the message is about.
The training dataset is composed by 810 samples,
with the following conspiracy label distribution: 435
Covid-Conspiracy, 242 Qanon-Conspiracy, 76 Flat
EarthConspiracy, 57 Pro-Russia Conspiracy. The hidden test
set is composed by 300 samples instead.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Models</title>
      <p>Due to the nature of the tasks, we mainly decided to try
diferent types of transformers based models for both
subtasks, in order to capture the semantics and the general
matter of the message itself. This is concatenated with
a densely connected neural network in order to classify
what the specific task is asking. Figure 1: The transformer architecture.</p>
      <p>
        More specifically a Transformer as described in
"Attention is all you need" [
        <xref ref-type="bibr" rid="ref3">10</xref>
        ], is composed of
encoderdecoder structure composed by multiple modules stacked 768. We executed fine tuning on the transformer.
ClasNx times on top of each other like in Figure 1 where each sification is executed on the first special output token
module is mainly consisted of Multi Head Attentions and [CLS] of the transformer
Feed Forward layers. In this architecture, the inputs and
the outputs (target sentences) are embedded (the outputs
need a right shift before usage) into an n-dimensional 4.2. XLM-RoBERTa
space because we cannot use the strings directly. XLM-RoBERTa [
        <xref ref-type="bibr" rid="ref19">25</xref>
        ] is a multilingual version of RoBERTa,
      </p>
      <p>
        Here we present the selected transformer-based mod- a transformers model pre-trained in a self-supervised
els for the tasks. Those were selected after a preliminary fashion, similarly to BERT, but with a larger corpus and
exploratory phase based on their performance on the no next sentecnce prediction. XLM-RoBERTa was
prevalidation set. trained on 2.5TB of filtered CommonCrawl data
containing 100 languages. Specifically, we used the
xlm-roberta4.1. BERT-xxl large variant, which has 24 hidden layers, 16 attention
heads and a hidden size of 1024. We executed fine tuning
We used the bert-base-italian-xxl-cased model[
        <xref ref-type="bibr" rid="ref18">24</xref>
        ], which on the transformer. Classification is executed on the first
is an Italian pretrained BERT, an encoder-only trans- special output token [CLS] of the transformer.
former, variant developed by MDZ Digital Library team.
      </p>
      <p>
        It was pretrained using as source data a Wikipedia dump
of various texts from the OPUS corpora collection with a 4.3. Llama
size of 13 GB and more than 2 billion tokens. With the LLaMA is an autoregressive language model developed
XXL variant, the corpus was extended with the Italian by Meta AI [
        <xref ref-type="bibr" rid="ref20">26</xref>
        ], based on a decoder only transformer
part of the OSCAR corpus, reaching a size of 81 GB and architecture. We used the 7B variant, the smallest one,
more than 13 billion tokens. This BERT-xxl model has which has 7 billions parameters. It was pretrained on
12 hidden layers, 12 attention heads and a hidden size of 1 trillion tokens from CCNet [67%], C4 [15%], GitHub
[4.5%], Wikipedia [4.5%], Books [4.5%], ArXiv [2.5%],
Stack Exchange[2%] sources. The Wikipedia and Books
sources are multilingual. Classification is executed on
the last output token of the transformer.
      </p>
      <p>We don’t use fine tuning on this model due to its size,
but only use it to generate sentence embeddings;
trainining was only executed on the classification head.</p>
      <sec id="sec-4-1">
        <title>4.4. Topic-specific tf-idf baseline</title>
        <p>For Subtask B, considered its nature of topic classification
and observing the presence of specific and unique words
in each topic, we also developed an original heuristic
baseline based on this assumptions. In short, it tries
to retrieve the most specific keywords to each topic and
extract their distribution in input texts. We recall that the
definition of tf-idf for each word  in a set of documents
 ∈  (in our case each document corresponds to each
Telegram message in the dataset) is:</p>
        <p>− , = , × 
, with , = , (, being the number of
occurren| |</p>
        <p>||
cies of word  in document ) and  = 10 :∈</p>
        <p>This method makes use of topic-specific tf-idf , which is
basically the normalized average tf-idf for each word in
respect to the documents of each topic, then divided by
the average tf-idf of the same word in the other topics.</p>
        <p>In mathematical terms, defining  as the set of topics,
_ , as the average tf-idf for word  and topic ,
and __ , as the normalized _ ,
in [0, 100] range, we have:
_ _ , =</p>
        <p>__ ,
∑︀′∈ ∖ __ ,′</p>
        <p>Instead, for the Topic-specific tf-idf model, as the focus
are topic specific relevant words, we apply stop word and
short words (less than 3 characters) removal, number and
punctuation elimination and stemming.</p>
        <p>This sccore is calculated only for the training
set; for each topic t then we extract the top K 5. Implementation
_ _ , words and store them (K is an
hyperparameter). Figure 2 shows the top 10 keywords We used the Python environment for developing the
modfor each topic with their respective score. els, using mainly PyTorch, Scikit-Learn and Transformers</p>
        <p>Finally, for each input text, we extract the distribu- libraries.
tion of the previously stored words, thus we obtain a
_ ×  distribution vector. This vector is then
fed into a Random Forest (RF) model for the final classifi- 6. Experiments and results
cation. This model is trained with 6-fold Cross-Validation
(CV) on the training set.</p>
        <p>We used an hold-out approach for both subtasks,
reserving 20% of the training set for validation for
hyperparameter tuning (split with labels ratio preservation). We
4.5. Preprocessing experimented with retrain on validation found
hyperpaFor the transformer-based models, only light preprocess- rameters, but with worse results, so we decided to keep
ing was applied, only substituting break line characters the model tested on the validation set as the final model
with spaces and using each transformer tokenizer. for each configuration.</p>
        <p>Model
BERT-xxl
RoBERTa-XLM
Llama 7B
Model
BERT-xxl
RoBERTa-XLM
Llama 7B
[1e-6, 2e-6, 3e-6]
[6e-6, 8e-6]
[1e-5, 5e-5, 1e-4, 5e-4]</p>
        <p>warmup</p>
        <p>For the Topic-specific tf-idf baseline, the validation cause of the benefits of finetuning or of the encoder-only
set was used for finding the best K. After this we used transformer architecture, versus the decoder and not
finea retrain strategy, in order to obtain a more general tuned Llama.
topic_specific_tfidf for words in each topic (RF classifier Among the relevant findings we include also that the
was also retrained with same CV hyperparameters) transformer dimension does not influence the
perfor</p>
        <p>The performance score of choice is macro-averaged F1 mance score; for example, although XLM-RoBERTa
emscore, as it is the one also used to evaluate the challenge. ploys a larger architecture than BERT, they are
comparable. The same reasoning applies when confronting with
6.1. Hyperparameters grid search Llama 7B, which has at least an order of magnitude more
parameters than the other transformers.</p>
        <p>Tables 1, 2 and 3 display the explored hyperparameters This indicates that the pre-training dataset (we recall
respectively for transformer-based models in SubtaskA, that BERT-xxl is not multilingual and trained only in
transformer-based models in SubtaskB and Topic-specific Italian) and the choice of finetuning have the greatest
tf-idf baseline model. The final chosen hyperparameters impact on performances.
are those which yield the best score on the validation set In regard to the Topic-specific tf-idf model, it provides
and are highlighted in bold. solid results in exchange for a lower computational cost,
thanks to its strong assumptions of the importance of
6.2. Results topic specific keywords in Subtask B.
It is also important to note that the samples correctly
Tables 4 and 5 display the scores on both the internal identified by Topic-specific tf-idf are not a strict subset
validation set (the score used to choose the model with of correctly identified samples by the BERT model, as
the best hyperparameters) and the hidden test set, respec- the predictions on the test set have a divergence ratio of
tively for SubtaskA and SubtaskB. Only macro-averaged almost 25%, while there is a performance diference of less
F1 score is reported in the tables. than 7%, meaning that a substantial set of "hard" (wrongly</p>
        <p>The whole hidden test set is split in public and private classified) samples for the transformer model are instead
test sets by the competition rules; the final test score is "easy" (correctly classified) for the Topic-specific tf-idf
obtained by weighted average (proportional each of the and vice versa. This implies that combining the 2 models
2 test set sizes) of the public and private sets. in a meaningful way could result in a more robust model.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>7. Discussions</title>
      <p>
        For both tasks, the best performing models are the
BERTbased ones, both the Italian BERT-xxl and XLM-RoBERTa,
as their performance is close in F1 terms and are the top-2
performers in both subtasks. These results are a probable
K
Random Forest max_depth
Random Forest max_features
Random Forest min_samples_leaf
Random Forest n_estimators
[
        <xref ref-type="bibr" rid="ref14 ref3">10, 20, 30, 40, 50, 60, 70, 80, 90, 100</xref>
        ]
[5, 15, None]
[log2, None]
[ [
        <xref ref-type="bibr" rid="ref1 ref5 ref7">1, 2, 4</xref>
        ]
[ [64, 128, 256]
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Menini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Polignano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sprugnoli</surname>
          </string-name>
          , G. Venturi,
          <year>Evalita 2023</year>
          :
          <article-title>Overview of the 8th evaluation campaign of natural language process</article-title>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Verginer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Casiraghi, BERT-xxl Spillover of antisocial behavior from fringe platXLMLla-RmoaB7EBRTa forms: The unintended consequences of community banning</article-title>
          ,
          <source>in: Proceedings of the International AAAI Conference on Web and Social Media</source>
          , volume
          <volume>17</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>742</fpage>
          -
          <lpage>753</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Casiraghi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Verginer</surname>
          </string-name>
          ,
          <article-title>Understanding online migration decisions followValidation score Test score ing the banning of radical communities</article-title>
          ,
          <source>arXiv BERT-xxl 0.8651 0</source>
          .8265 preprint arXiv:
          <volume>2212</volume>
          .04765 (
          <year>2022</year>
          ).
          <source>XLMLla-RmoaB7EBRTa 00..88717263</source>
          <volume>00</volume>
          ..
          <volume>87538392</volume>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Topic-specific
          <source>tf-idf 0.7400 0</source>
          .7520 Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          , Attention is all you need,
          <year>2017</year>
          . arXiv:
          <volume>1706</volume>
          .
          <fpage>03762</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gaut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>ElSherief, ing and speech tools for italian</article-title>
          , in: Proceedings J.
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Mirza</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Belding</surname>
            ,
            <given-names>K.-W.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , W. Y.
          <article-title>of the Eighth Evaluation Campaign of Natural Lan- Wang, Mitigating gender bias in natural language guage Processing and Speech Tools for Italian. Final processing: Literature review</article-title>
          ,
          <source>arXiv preprint Workshop (EVALITA</source>
          <year>2023</year>
          ), CEUR.org, Parma, Italy, arXiv:
          <year>1906</year>
          .
          <volume>08976</volume>
          (
          <year>2019</year>
          ).
          <year>2023</year>
          . [12]
          <string-name>
            <given-names>G.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hollenstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Musat</surname>
          </string-name>
          , C. Zhang,
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stoehr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <article-title>Acti at evalita Control, generate, augment: A scalable framework 2023: Overview of the conspiracy theory identifica- for multi-attribute text generation, in: Findings tion task</article-title>
          ,
          <source>arXiv preprint arXiv:2307.06954</source>
          (
          <year>2023</year>
          ).
          <article-title>of the Association for Computational Linguistics:</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H. W.</given-names>
            <surname>Hanley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Durumeric</surname>
          </string-name>
          ,
          <source>A golden EMNLP</source>
          <year>2020</year>
          ,
          <article-title>Association for Computational Linage: Conspiracy theories' relationship with mis- guistics,</article-title>
          <string-name>
            <surname>Online</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>351</fpage>
          -
          <lpage>366</lpage>
          . URL:
          <article-title>https: information outlets, news media, and the wider //aclanthology</article-title>
          .org/
          <year>2020</year>
          .findings- emnlp.33. internet (preprint) (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .findings-emnlp.
          <volume>33</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Chandrasekharan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Pavalanathan</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Srini- [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gote</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Brandenberger</surname>
          </string-name>
          , S. Schlosser, vasan, A. Glynn,
          <string-name>
            <given-names>J.</given-names>
            <surname>Eisenstein</surname>
          </string-name>
          , E. Gilbert, You F.
          <article-title>Schweitzer, Disentangling active and pascan't stay here: The eficacy of reddit's 2015 ban sive cosponsorship in the u.s. congress, ArXiv examined through hate speech</article-title>
          ,
          <source>Proc. ACM Hum.- abs/2205</source>
          .09674 (
          <year>2022</year>
          ).
          <source>Comput. Interact</source>
          .
          <volume>1</volume>
          (
          <year>2017</year>
          ). URL: https://doi.org/10 [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Valvoda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pimentel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stoehr</surname>
          </string-name>
          , R. Cotterell, .
          <volume>1145</volume>
          /3134666. doi:
          <volume>10</volume>
          .1145/3134666. S. Teufel,
          <source>What about the precedent: An</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>E.</given-names>
            <surname>Chandrasekharan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jhaver</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Bruckman, information
          <article-title>-theoretic analysis of common law</article-title>
          , in: E. Gilbert,
          <article-title>Quarantined! examining the efects Proceedings of the 2021 Conference of the North of a community-wide moderation intervention on American Chapter of the Association for Computareddit, ACM Transactions on Computer-Human tional Linguistics: Human Language Technologies, Interaction (TOCHI) 29 (</article-title>
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          . Association for Computational Linguistics, Online,
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Trujillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cresci</surname>
          </string-name>
          , Make reddit great again: as- 2021, pp.
          <fpage>2275</fpage>
          -
          <lpage>2288</lpage>
          . URL: https://aclanthology.org sessing community efects of moderation interven- /
          <year>2021</year>
          .naacl-main.
          <volume>181</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .n tions on r/the_donald,
          <source>Proceedings of the ACM on aacl-main.181. Human-Computer Interaction</source>
          <volume>6</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
          . [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dhuliawala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stoehr</surname>
          </string-name>
          ,
          <article-title>Extracting victim counts from text</article-title>
          ,
          <source>arXiv preprint arXiv:2302.12367</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Alonso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Saini</surname>
          </string-name>
          , G. Kovács,
          <article-title>Hate speech detection using transformer ensembles on the hasoc dataset</article-title>
          , in: Speech and Computer: 22nd International Conference, SPECOM 2020,
          <article-title>St</article-title>
          . Petersburg, Russia, October 7-
          <issue>9</issue>
          ,
          <year>2020</year>
          , Proceedings, Springer,
          <year>2020</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R. T.</given-names>
            <surname>Mutanga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Naicker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. O.</given-names>
            <surname>Olugbara</surname>
          </string-name>
          ,
          <article-title>Hate speech detection in twitter using transformer methods</article-title>
          ,
          <source>International Journal of Advanced Computer Science and Applications</source>
          <volume>11</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>L.</given-names>
            <surname>Stappen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Brunn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schuller</surname>
          </string-name>
          ,
          <article-title>Cross-lingual zero-and few-shot hate speech detection utilising frozen transformer language models and axel</article-title>
          , arXiv preprint arXiv:
          <year>2004</year>
          .
          <volume>13850</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Louf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , P. von Platen, C. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gugger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Drame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lhoest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Rush</surname>
          </string-name>
          ,
          <article-title>Huggingface's transformers: State-of-the-art natural language processing</article-title>
          ,
          <year>2020</year>
          . arXiv:
          <year>1910</year>
          .03771.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <year>2019</year>
          . arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>C.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>How to ifne-tune bert for text classification</article-title>
          ?,
          <year>2020</year>
          . arXiv:
          <year>1905</year>
          .05583.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>H.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chung</surname>
          </string-name>
          , G. Friedland,
          <article-title>Detecting conspiracy theories from tweets: Textual and structural approaches</article-title>
          ., in: MediaEval,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>W.</given-names>
            <surname>Marcellino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. C.</given-names>
            <surname>Helmus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kerrigan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Reininger</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. I. Karimov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Lawrence</surname>
          </string-name>
          ,
          <article-title>Detecting Conspiracy Theories on Social Media: Improving Machine Learning to Detect and Understand Online Conspiracy Theories</article-title>
          , RAND Corporation, Santa Monica, CA,
          <year>2021</year>
          . doi:
          <volume>10</volume>
          .7249/RR-A676
          <article-title>-1.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [24]
          <article-title>DBMDZ, bert-base-italian-xxl-</article-title>
          <string-name>
            <surname>cased</surname>
          </string-name>
          ,
          <year>2020</year>
          . URL: ht tps://huggingface.co/dbmdz/bert-base
          <article-title>-italian-xxl -cased.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Unsupervised crosslingual representation learning at scale</article-title>
          ,
          <year>2020</year>
          . arXiv:
          <year>1911</year>
          .02116.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample,
          <article-title>Llama: Open and eficient foundation language models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>