<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Tackling Covid-19 Conspiracies on Twitter using BERT Ensembles, GPT-3 Augmentation, and Graph NNs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Damir Korenčić</string-name>
          <email>damir.korencic@irb.hr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Grubišić</string-name>
          <email>ivan.grubisic@irb.hr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gretel Liz De La Peña Sarracén</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alejandro Hector Toselli</string-name>
          <email>ahector@prhlt.upv.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Berta Chulvi</string-name>
          <email>berta.chulvi@upv.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Paolo Rosso</string-name>
          <email>prosso@dsic.upv.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ruđer Bošković Institute</institution>
          ,
          <addr-line>Zagreb</addr-line>
          ,
          <country country="HR">Croatia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We describe several approaches to text- and graph-based classification for detecting COVID-19 conspiracies on Twitter. We tackle the tasks of text classification with and without graph data, and classification of Twitter users based on user graph. To this end, we experiment with large transformer ensembles, GPT-3-based techniques, and a variety of graph neural networks. Our results demonstrate that transformer ensembling and GPT-3 text augmentation can improve performance and stability, and that richer graph data does not necessarily lead to improved performance.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Motivation and Background</title>
      <p>
        The COVID-19 pandemic has increased the amount of people’s exposure to digital content and
communications [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. One of the efects of the rapid digitalization of everyday activities is the
increase of online disinformation, including conspiracy theories [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. A conspiracy theory can
be defined as a belief that two or more actors have coordinated in secret to achieve a malevolent
outcome [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The automatic detection of conspiracy messages on social networks has a huge
potential for combating disinformation.
      </p>
      <p>
        MediaEval 2022 FakeNews challenge consists of three subtasks focused on detecting
COVID19 conspiracy theories in tweets. Nine diferent conspiracy categories are defined. Subtask 1
consists of determining, for each conspiracy theory, whether a tweet supports it, mentions it, or
ignores it. In Subtask 2 the goal is to determine, based on a social network graph, whether a
Twitter user is a misinformation poster. The goal of Subtask 3 is the same as that of Subtask 1,
but the model can use the tweet’s user ID and all the graph data. Both the text and the user
datasets contain 1,913 labeled development tweets/users and 830 test tweets/users. In addition,
a large user graph, with 1,679,011 vertices and 268,694,698 edges, is provided. The metric of
the models’ performance assessment is Matthews correlation coeficient (MCC) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. A detailed
description of the subtasks and the datasets can be found in Pogorelov et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Our approach to Subtask 1 was to improve on the best-performing transformer-based system
of MediaEval 2021 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] by adding additional features, creating model ensembles, and augmenting
the training data via GPT-3 model. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. For subtasks 2 and 3 we designed and evaluated a
number of GNN models [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], each consisting of two convolutional layers, three fully connected
layers, and a softmax classification layer.
      </p>
      <p>
        Few works have addressed the detection of conspiracy theories. These include the
classiifcation models proposed in the MediaEval 2021 challenge [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and in the work of Giachanou
et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. There exist numerous approaches to text classification [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and in recent years deep
learning models and especially transformers have led to state-of-art results [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Graph neural
networks (GNNs) have been widely used in recent years for diferent classification tasks [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ],
including text classification [
        <xref ref-type="bibr" rid="ref15 ref16 ref17">15, 16, 17</xref>
        ] where they can lead to top results [
        <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
        ]. The strength
of GNNs is their ability to model relationships between the data points and integrate them with
other features.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Approach to Text-based Classification</title>
      <p>
        Subtask 1 of MedEval FakeNews 2022 is identical to Subtask 3 of MedEval 2021, but with a
larger dataset. Our approach is therefore to improve on the best system from the 2021 challenge
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] (hard baseline), based on domain-specific COVID-Twitter-BERT (CT-BERT) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] fine-tuned
in a multi-task manner. We attempted to append additional features (TF-IDF, topic models, and
hand-crafted features) before applying the classification layer, but this led to no improvement.
We also experimented with diferent architectures for the multi-task classifier layer (attention,
additional layers), but the basic logistic regression gave the best results.
      </p>
      <p>We improved upon the hard baseline model by combining a large number of transformers
and examining diferent options for ensemble creation. This approach is motivated by the
observation that model performances vary significantly, in relation to both the eval/dev/train
split and random initialization. Individual models all use the same preprocessing and training
parameters and difer only in random seeds and train/dev data splits.</p>
      <p>
        Preprocessing and learning settings were tweaked in the initial development phases of the
experiment. We tried the text preprocessing options from [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] but opted to use no preprocessing
for the final models. We used the Adam optimizer with a weight decay of 10− 3 and an initial
learning rate (LR) of 7 * 10− 5. We reduced the LR by a factor of 0.3 after 3 epochs of no
improvement and stop the training after 6 such epochs. LR used for fine-tuning the transformer
parameters is 10× smaller. We used a nested evaluation scheme based on stratified
crossvalidation. On the first level, 7-fold splitting is performed, followed by the second level 6-fold
splitting. This way the development data is split into eval (14.3%), dev (14.3%), and train
(71.4%) subsets.
      </p>
      <p>
        We also attempted to use the GPT-3 (davinci) model, by way of in-context learning [
        <xref ref-type="bibr" rid="ref19 ref7">7, 19</xref>
        ].
This approach proved inefective, but a simple data augmentation method – the addition of
GPT-rephrased tweets with original labels – yielded competitive results. Additionally, we
experimented with the Xgboost classifier [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] applied to topic model features.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Approach to Graph-based Classification</title>
      <p>Graph-based User Classification (Subtask 2) Subtask 2 consists of classifying Twitter
users as (not) being misinformation spreaders, and we tackle it using graph neural networks
(GNNs). The graph nodes are described by a matrix  where each row is a vector containing a
single user’s attributes (location, number of followers, . . . ). Note that our models do not take
into account the user’s textual information, in order to examine the viability of using graph-only
data. Edges are represented with a square matrix  – each element corresponds to the weight
of the relation between two users (the weight being 0 if there is no relation).</p>
      <p>
        GNN classification models consist of two convolutional layers, followed by three fully
connected layers of 150 neurons (each followed by a normalization layer), followed by the standard
softmax classification layer. We have tested 16 combinations of diferent convolutional layers –
GCNConv [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], GATv2Conv [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], TAGConv [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and RGGConv [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. We trained the model in
200 epochs with a learning rate of 0.01 and a batch size of 64.
      </p>
      <sec id="sec-3-1">
        <title>Graph-based Text Classification (Subtask 3) Subtask 3 has the same goal as Subtask 1 –</title>
        <p>classifying texts for each of the 9 conspiracy categories. However, in Subtask 3 the information
about user connectivity can be added to the classifiers. We used the same approach as for
Subtask 2, but with a modified graph-construction method.</p>
        <p>In this subtask, the nodes are texts (instead of users) and each node is represented with a
vector of features extracted from a text using pretrained transformer models. Optionally, we
added the vector of user attributes describing the text’s author. An edge exists between two text
nodes if the authors of the corresponding texts are related, and the edge weight is derived as in
Subtask 2. Another diference with Subtask 2 is that here we solve a multi-label and multi-class
classification problem (instead of binary classification). We trained a classifier per conspiracy
category and combined the results to obtain the final output.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Analysis of the Text-based Classification</title>
      <p>Best-performing models were selected by varying several options for scoring and ensemble
creation. The combinations were evaluated on eval/dev/train splits – each outer eval split
was used to calculate the performance of models and ensembles created from inner splits. For
each inner split, we trained 40 models (1680 models in total), changing only the random seed
for initializing the classification layer parameters.</p>
      <p>For each inner dev/train split, the best model from all the epochs was selected according
to a scoring method (score option) – either by selecting a model with top average MCC, or the
top model per conspiracy category. Optionally, a number of top models (up to 15) were selected
(#models option). All the models of inner splits were aggregated, either by majority voting
(vote) or by multiplication of the models’ class probabilities (mult). For each combination of
options, all the models from inner splits are aggregated and the performances of these ensembles
on eval splits are averaged to obtain the performance assessment for model selection. For
the final submission, we selected the best ensemble configurations (models 2–4) and used the
average of eval, dev, and train split scores for inner fold model selection. We also created
and evaluated a “backup” model (model 1) created with the default options and using only eval
splits for inner model scoring. The ensemble of models trained with the GPT3-augmented data
(model 5) was created in the same manner as the backup model. Augmentation was performed
by rephrasing the tweets of the entire development set and using them (labeled with original
labels) to duplicate the size of each train split. The best version of the Xgboost classifier
(model 7) uses topic models and SVD features.</p>
      <p>The results are presented in Table 1. It can be seen that ensembling improves the
singlemodel solution (model 0), by increasing the expected mean performance and the worst-case
performance. Interestingly, GPT-3 augmentation without ensembling (model 6) has a similar
efect. All the deep ensemble variants perform very similarly and have good generalization in
terms of dev/test score diferences.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Analysis of the Graph-based Classification</title>
      <sec id="sec-5-1">
        <title>Graph-based User Classification (Subtask 2) In the construction of the model we consider</title>
        <p>4 diferent convolutional layers described in Section 3 (GCN, GATv2, TAG, and RGGConv).
We evaluated all 16 two-layer combinations to discover the best design for the task. We also
considered two types of graphs, one with all the users in the dataset (big) and another with only
the users from dev and train sets of Subtask 2 (small).</p>
      </sec>
      <sec id="sec-5-2">
        <title>Graph-based Text Classification (Subtask 3) For Subtask 3 we evaluated combinations</title>
        <p>of two node representation methods: text embeddings and user attributes. We evaluated two
models to obtain the embeddings: COVID-Twitter-BERT-v2 (covid-bert) and all-MiniLM-L6-v2
(miniLM). The results in Table 2 indicate the importance of the node representation – the MCC
values dropped steeply after user attributes were added to text embeddings, and the embeddings
obtained with the COVID-Twitter-BERT-v2 model seem to perform better.</p>
        <p>Additionally, we submitted text-only solutions described in Section 4 – two best transformer
ensembles and the Xgboost classifier. The two ensembles have test MCCs of 0.698 and 0.697,
while the Xgboost classifier has a test MCC of 0.395. This indicates that the graph information
might be detrimental to text classification.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>
        The results for Subtask 1 show that both large transformer ensembles and simple GPT-3 text
augmentation can improve the performance of individual models and lower the expected
performance variance. However, the performance of the best models is similar and it seems
to have reached a limit of cca. 0.74 MCC. This could be due to the inherent task complexity,
or to the noise present in existing labels. It is interesting that the “in-context learning” with
GPT-3, which tends to produce decent results [
        <xref ref-type="bibr" rid="ref19 ref7">7, 19</xref>
        ], failed for this use case (with 0.172 MCC).
We hypothesize that this is due to inherent task complexity: GPT-3 cannot distinguish, based
on complex prompts with category descriptions, between fine-grained conspiracy categories.
      </p>
      <p>Acknowledgments: This work was carried out in the framework of the following projects:
XAI-Disinfodemics (PLEC2021-007681), IBERIFIER (INEA/CEF/ICT/A202072381931, n.
2020-EUIA-0252), and MARTINI CHIST-ERA (PCI2022-134990-2). It was also supported by the Maria
Zambrano grant (Spanish Ministerio de Universidades, NextGenerationEU/PRTR).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>European</given-names>
            <surname>Commission</surname>
          </string-name>
          ,
          <source>Digital economy and society index (desi)</source>
          <year>2022</year>
          ,
          <year>2022</year>
          . URL: https:// digital-strategy.ec.europa.eu/en/policies/desi.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B.</given-names>
            <surname>European Commission</surname>
          </string-name>
          ,
          <article-title>Flash eurobarometer 464 (fake news and disinformation online)</article-title>
          ,
          <source>GESIS Data Archive, Cologne. ZA6934 Data file Version 1.0.0</source>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .4232/1.13019.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>K. M. Douglas</surname>
            ,
            <given-names>M. S.</given-names>
          </string-name>
          <string-name>
            <surname>Robbie</surname>
          </string-name>
          ,
          <article-title>What are conspiracy theories? a definitional approach to their correlates, consequences, and communication, Annual review of psychology (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gorodkin</surname>
          </string-name>
          ,
          <article-title>Comparing two k-category assignments by a k-category correlation coeficient</article-title>
          ,
          <source>Computational biology and chemistry 28</source>
          (
          <year>2004</year>
          )
          <fpage>367</fpage>
          -
          <lpage>374</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pogorelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. T.</given-names>
            <surname>Schroeder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Brenner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Maulana</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Langguth,</surname>
          </string-name>
          <article-title>Combining tweets and connections graph for fakenews detection at mediaeval 2022</article-title>
          ,
          <source>in: Proceedings of the MediaEval 2022 Workshop</source>
          , Bergen, Norway and Online,
          <volume>12</volume>
          -
          <fpage>13</fpage>
          January
          <year>2023</year>
          .,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peskine</surname>
          </string-name>
          , G. Alfarano,
          <string-name>
            <surname>I. Harrando</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Troncy</surname>
          </string-name>
          , Detecting covid-19
          <article-title>-related conspiracy theories in tweets</article-title>
          ,
          <source>in: MediaEval</source>
          <year>2021</year>
          ,
          <source>MediaEval Benchmarking Initiative for Multimedia Evaluation Workshop</source>
          ,
          <fpage>13</fpage>
          -15
          <source>December</source>
          <year>2021</year>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          , et al.,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Y.</given-names>
            <surname>Philip</surname>
          </string-name>
          ,
          <article-title>A comprehensive survey on graph neural networks</article-title>
          ,
          <source>IEEE transactions on neural networks and learning systems 32</source>
          (
          <year>2020</year>
          )
          <fpage>4</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pogorelov</surname>
          </string-name>
          , D. T. Schroeder13,
          <string-name>
            <given-names>S.</given-names>
            <surname>Brenner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Langguth</surname>
          </string-name>
          ,
          <article-title>Fakenews: Corona virus and conspiracies multimedia analysis task at mediaeval 2021</article-title>
          , in: Multimedia Benchmark Workshop,
          <year>2021</year>
          , p.
          <fpage>67</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Giachanou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ghanem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <article-title>Detection of conspiracy propagators using psycho-linguistic characteristics</article-title>
          ,
          <source>Journal of Information Science</source>
          (
          <year>2021</year>
          ). doi:0165551520985486.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kowsari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jafari Meimandi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Heidarysafa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mendu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Barnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Brown</surname>
          </string-name>
          , Text Classification Algorithms: A Survey,
          <source>Information</source>
          <volume>10</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .3390/info10040150.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Minaee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kalchbrenner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Cambria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nikzad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chenaghlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <article-title>Deep Learning-based Text Classification: A Comprehensive Review</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>54</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>40</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3439726. doi:
          <volume>10</volume>
          .1145/3439726.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , G. Cui,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Graph neural networks: A review of methods and applications</article-title>
          ,
          <source>AI</source>
          Open 1
          <article-title>(</article-title>
          <year>2020</year>
          )
          <fpage>57</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sato</surname>
          </string-name>
          ,
          <article-title>A survey on the expressive power of graph neural networks</article-title>
          , arXiv:
          <year>2003</year>
          .
          <volume>04078</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          , S. C. Han,
          <string-name>
            <surname>S</surname>
          </string-name>
          . Long,
          <string-name>
            <given-names>J.</given-names>
            <surname>Poon</surname>
          </string-name>
          ,
          <article-title>Me-gcn: Multi-dimensional edge-embedded graph convolutional networks for semi-supervised text classification</article-title>
          ,
          <source>arXiv preprint arXiv:2204.04618</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>G. L. De la Peña Sarracén</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
          </string-name>
          ,
          <article-title>Unsupervised embeddings with graph auto-encoders for multi-domain and multilingual hate speech detection</article-title>
          ,
          <source>in: Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>2196</fpage>
          -
          <lpage>2204</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>G. L. De la Peña Sarracén</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
          </string-name>
          ,
          <article-title>Convolutional graph neural networks for hate speech detection in data-poor settings</article-title>
          ,
          <source>in: International Conference on Applications of Natural Language to Information Systems</source>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Salathé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Kummervold</surname>
          </string-name>
          ,
          <article-title>Covid-twitter-bert: A natural language processing model to analyse covid-19 content on twitter</article-title>
          , arXiv preprint arXiv:
          <year>2005</year>
          .
          <volume>07503</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Ornstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. N.</given-names>
            <surname>Blasingame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Truscott</surname>
          </string-name>
          , How To Train Your Stochastic Parrot:
          <article-title>Deep Language Models for Political Texts (</article-title>
          <year>2022</year>
          )
          <fpage>30</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Benesty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Khotilovich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <article-title>Xgboost: extreme gradient boosting</article-title>
          ,
          <source>R package version 0.4-2 1</source>
          (
          <issue>2015</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Kipf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <article-title>Semi-supervised classification with graph convolutional networks</article-title>
          ,
          <source>CoRR abs/1609</source>
          .02907 (
          <year>2016</year>
          ). arXiv:
          <volume>1609</volume>
          .
          <fpage>02907</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S.</given-names>
            <surname>Brody</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Alon</surname>
          </string-name>
          , E. Yahav,
          <article-title>How attentive are graph attention networks?</article-title>
          ,
          <source>CoRR abs/2105</source>
          .14491 (
          <year>2021</year>
          ). arXiv:
          <volume>2105</volume>
          .
          <fpage>14491</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Wu,
          <string-name>
            <given-names>J. M. F.</given-names>
            <surname>Moura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kar</surname>
          </string-name>
          ,
          <article-title>Topology adaptive graph convolutional networks</article-title>
          ,
          <source>CoRR abs/1710</source>
          .10370 (
          <year>2017</year>
          ). arXiv:
          <volume>1710</volume>
          .
          <fpage>10370</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>X.</given-names>
            <surname>Bresson</surname>
          </string-name>
          , T. Laurent,
          <article-title>Residual gated graph convnets</article-title>
          ,
          <source>CoRR abs/1711</source>
          .07553 (
          <year>2017</year>
          ). arXiv:
          <volume>1711</volume>
          .
          <fpage>07553</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>