<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Reducing Misinformation and Toxic Content Using Cross-Lingual Text Summarization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hoai Nam Tran</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Udo Kruschwitz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Regensburg</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Misinformation has long been considered a major problem in our digital world, but automatically identifying it still remains a challenging issue. It becomes even more of a problem when tackling content written in languages other than English. We also note that much progress has been made in classifying short social media posts, but there are many other types of misinformation. We present steps towards addressing the problem by adopting ideas that have shown to be promising in related prior work, namely applying extractive and abstractive text summarization methods so that we can process documents of any length and by incorporating machine translation as part of our overall architecture. We consider misinformation as just one out of many types of content that should be identified automatically on the way to a healthier digital ecosystem and see toxic content such as hate speech as naturally falling within the same scope of our work. We demonstrate on several benchmark collections covering both misinformation and toxic content that our approach is robust and achieves competitive performance on these datasets. This ofers plenty of scope for future work. To foster reproducibility, we make all code and models available to the community via GitHub and Hugging Face.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Misinformation</kwd>
        <kwd>Text summarization</kwd>
        <kwd>Toxic content detection</kwd>
        <kwd>Cross-lingual</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Fake News and Hate Speech have one thing in common: the aim is to spread toxicity and to
bring harm to the world. They have now become serious and significant social and political
issues [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. How did we get there? One aspect is that users have been shown to get more
easily persuaded and influenced by social media posts, causing them to change their attitude
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. In combination with the excessive usage of social media, the desire for validation and the
fear of rejection negatively impact our mental health, especially for teenagers and children
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Much progress has been made recently in addressing the challenge, often focussing on
social media. Searching for relevant information with common information retrieval systems
and natural language processing pipelines gets more complicated with the amount of harmful
misinformation. The flood of toxic content and polarization leads to distrust in any news
channel; e.g., only 26% of American adults trust any news media [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], which is why we need
to improve the quality of the information we consume. Most competitive approaches include
incorporating transformer-based models like BERT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] to assist social media moderators and
fact-checkers in combating harmful content. However, since news articles and popular blog
posts are also afected by misinformation and hateful assertions, and transformer-based models
have a limited input size (e.g., 512 for BERT), the challenge here is to find a way to also use
these models efectively for longer texts. This is where we propose text summarization. The
second motivation is the fact that very few languages can be considered resource-rich, making
it desirable to tap into such resources when tackling toxic content in other languages.
      </p>
      <p>This paper presents a framework combining automatic machine translation, text
summarization, and classification, tackling misinformation and toxic content. We provide experimental
results for several common benchmarks using both binary and multi-class classification. To
foster reproducibility, we make all code, hyperparameters, and detailed result tables available
via GitHub1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>This section provides an overview of related work in fake news detection, hate speech detection,
multilingual machine translation, and text summarization. Since misinformation leads to toxic
polarization, which again leads to abusive language, and hate speech is generally considered
harmful, we consider both fields to fall within the scope of detecting toxic content.</p>
      <p>
        Fake news detection and hate speech detection (HSD) are two active research areas,
with research often guided by shared tasks and competitions, e.g., as part of CLEF, SemEval,
or GermEval. While users usually try to write more engaging comments to achieve more user
interactions, the user’s “dark side” [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] is the posting of hateful comments, including toxic and
ofensive language. There are monolingual and multilingual approaches to detect hate speech
and toxic comments since online comments can be written in diferent languages and possibly
a mix of several. A number of diferent approaches can be adopted to tackle this task, and at a
high level one can distinguish content-based and context-based methods [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. We are focusing on
content-based approaches. Rather than providing a review of this massive body of literature we
just point out that Transformer models with self-attention [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] like BERT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], BART [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and
T5 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] dominate the field. HateBERT [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] is a BERT variant pre-trained with abusive online
community data from the social news and discussion platform Reddit2. A provided list of criteria
from tweets as predictive features can help to identify racist and sexist insults [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. They are
all Twitter-based in the HSD domain, similar to some of the datasets we use for our approach.
A survey of datasets on the topic of fake news detection and fact verification includes several
fact-checking sites [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Full Fact3 is an example of a fact-checking organisation that aims at
identifying harmful content with intelligence and monitoring tools, e.g., CrowdTangle, which
helps the user with manual fact-claim checking by raising alerts if exact user-defined keywords
are triggered [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Multilinguality is commonly addressed by using transformer models trained
on multiple languages, e.g., XLM-RoBERTa [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] has cross-lingual capabilities to work in several
tasks and benchmarks containing harmful texts which can appear in multiple languages. Fusion
1https://github.com/HN-Tran/ROMCIR_2023
2https://www.reddit.com/
3https://fullfact.org/
strategies with mBERT and XLM-RoBERTa for multilingual toxic text detection [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] or deep
learning ensembles for efective hate speech detection are other approaches that are similar to
ours. Given that multilingual models still focus on a limited number of languages for pre-training,
we explore automatic machine translation as an alternative.
      </p>
      <p>
        Machine translation is an essential part of many online services nowadays. In a survey
by CSA Research, 76% of online shoppers prefer information in their native language, and
40% would never buy from websites with only other languages [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Also, the global machine
translation market has increased from 450 million USD in 2017 to 1.1 billion USD in 2022 [
        <xref ref-type="bibr" rid="ref19 ref20">19, 20</xref>
        ].
The two most popular translation services are Google Translate [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] and Microsoft Translator
[22]. Both services are available in multiple languages and can be used for free. DeepL Translator
is a relatively new translation service that uses proprietary neural networks to translate text
[23]. They claim to surpass Google Translate and Microsoft Translator in terms of quality and
speed in several European languages [24]. These services are however not always accurate
and can even be exploited by malicious actors [25, 26]. Since transformer-based models and
automatic translation services are limited in their input length, summarization is our approach
to overcome this limitation.
      </p>
      <p>Summarization is still an active research field that successfully utilizes extractive machine
learning [27, 28] and abstractive approaches [29, 30, 31]. It has only recently been considered
in this context with state-of-the-art performance for fake news detection using a common
reference benchmark collection reported [32]. We propose to utilize progress in the field by
providing a novel combination of established techniques, leaving plenty of room for future
work to explore this idea further.</p>
      <p>In summary, we observe that misinformation and toxic content detection are conceptually
related areas that remain open problems despite progress that has been made in recent years.
We are interested in exploring content-based ideas that have shown promise in previous work
and see our contribution as one possible direction that utilises diferent types of automatic text
summarisation as well as machine translation. Future work can then explore this in more depth
and breadth.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>Our general framework is a pipeline-based architecture, as illustrated in Figure 1. It has three
main components: automatic machine translation, summarization, and classification. Due
to the availability of common benchmarks in German, it is our language of choice for the
source documents (but we obviously envisage this approach to be applied to actually
underresourced languages in future work). Each dataset gets machine-translated into English which
is followed by transformer-based text summarization. German-based models take the original
texts/comments and summarized texts as the input in the fine-tuning process, while
domainspecific and multilingual models use the translations. We train our models 5 times (runs) with
a diferent seed, where the model with the highest macro F1 score in each 50 steps for each
run gets chosen, and the inference outputs the predictions of each model. After that, all the 5
runs get ensembled in both majority voting types (hard and soft voting). Finally, the ensembled
models are ensembled again. Finding the optimal number of models for the ensemble is dificult
German
comments</p>
      <p>Labels</p>
      <p>Machine
Translation</p>
      <p>Summarization</p>
      <p>Transformers (Trainer)</p>
      <p>English models
Multilingual models
German models</p>
      <p>Ensembles
Ensembles
Ensembles</p>
      <p>Ensemble
strategy
because of the danger of overfitting and averaging. Thus we set a fixed number of 5 runs.</p>
      <p>German tasks and resources for training are sparse since English is the most represented
language in many benchmarks (~1000 tasks). German (~30 tasks) is even less represented than
other languages like Spanish (~60 tasks), Hindi (~45 tasks), and even Bengali (~35 tasks) [33].
Thus, we have decided on bilingual German and English datasets that can later be applied to
other languages.</p>
      <p>For fine-tuning, we use the recommended arithmetic mean over harmonic mean (see Equation
1) due to its robustness towards error type distribution [34]:
ℱ1 = 1 ∑︁ F1 = 1 ∑︁ 2</p>
      <p>+ 
2 ¯ ¯ ( 1 ∑︀ )( 1 ∑︀ )</p>
      <p>F1 = ( ¯ , ¯) = ¯ + ¯ = 2 1∑︀  + 1 ∑︀ 
As the GermEval metric, we use the harmonic mean over arithmetic mean (see Equation 2):
(1)
(2)</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>4.1. Data
Here we will briefly outline the experimental setup, choice of tools, and datasets.
We use the following shared task datasets (there is clearly scope to explore other, less-resourced
languages and classification tasks in future work):
• GermEval 2018 Subtask 1 [35]
• GermEval 2019 Task 2 Subtask 1 [36]
• GermEval 2021 Subtasks 1-3 [37]
• CLEF 2022 CheckThat! Lab Task 3 [38]</p>
      <p>All datasets contain an imbalance in their class label distributions (see Table 1), and the
number of characters are also very diferent. GermEval datasets fit the short text scenario since
they are comments from social networks and the CheckThat! dataset fits the long text scenario
with a maximum size of 100,000 characters (see Table 3). Since BERT models usually have a
maximum token limit of 512, the input would automatically be truncated to the first 512 tokens,
and thus relevant information might get lost in this process.</p>
      <sec id="sec-4-1">
        <title>4.2. Automatic Machine Translation</title>
        <p>For machine translation, we can choose between a text generation model like T5 [39] or
commercial translation services for our purpose. Since the translation quality depends on</p>
        <p>GBERT
GELECTRA
5e-6</p>
        <p>XLM-R
BERTweet</p>
        <p>1e-5
GermEval</p>
        <p>CLEF CheckThat! 2022
BERT-based</p>
        <p>T5-based BERT-based T5-based
2e-5
its pre-trained corpora quantity and quality, we decided on the two most popular machine
translation services: Google Translate and DeepL Translator.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Splitting Methods</title>
        <p>There are two standard options for deciding how to split our data for the fine-tuning process:
Fixed random seeds and k-fold cross-validation. In [40], they used random seeding, and since
we made five runs with an imbalanced dataset, a stratified 5-fold cross-validation was the other
option. Our results show that using fixed random seed values is better than using stratified
k-fold cross-validation.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.4. Hyperparameter Optimization</title>
        <p>Since searching for the optimal hyperparameters for our models is dificult, especially looking
for ways to avoid overfitting, we use the Optuna [ 41] library, which can be integrated into the
Hugging Face Trainer library as an option for hyperparameter search. Since even the default
hyperparameters can lead to overfitting in specific benchmark datasets, the chance of having
similar data points between the development and test set is given. We tested 100 combinations,
which evaluates the best possible setting for our macro-F1 metric (see Equation 1). The question
is whether a complete automatic hyperparameter search can be conducted by a tool like Optuna
to work efectively without looking for any working hyperparameters.</p>
        <p>After choosing the first three best runs, the results show that it is an appropriate way to
ifnd parameters for the development set but not for the test set. We have not tested this on
a fixed amount of known default hyperparameters yet. Since hyperparameter optimization
is also a very time-consuming process, we have decided to use each model’s recommended
hyperparameters if reported in the corresponding papers. If not, we use the exact parameters of
their model architecture.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.5. Summarization</title>
        <p>We use both extractive and abstractive summarization separately and exclusively. Since we only
have one textual input, we first concatenate the title and the text with a dot so that the title is
considered the first sentence (see Equation 3). Sometimes, the title is written like clickbait, a
sentence without any information-relevant value.</p>
        <p>+ .</p>
        <p>+ 
(3)</p>
        <p>
          Extraction-based summarization aims to select the most relevant representations of the
given text input. In the used library [42], we apply k-means clustering and use the Elbow
method to find the optimal  [43, 44]. Our chosen model is DistilBART-CNN-12-64 which is
based on BART [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] with distillation [45], fine-tuned with the CNN and DailyMail dataset [ 46].
        </p>
        <p>
          Abstraction-based summarization aims to generate shorter text with the most relevant
representations of the given text input. We use the version of T5 [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] with three billion
parameters (T5-3B) to generate shorter text with the identical prompt template ("summarize:")
used in the pre-training process for the CNN/DailyMail dataset [46]. With the use of relative
positional embeddings, the utilization of much longer text at the cost of higher computing
consumption is possible [
          <xref ref-type="bibr" rid="ref11">47, 11</xref>
          ].
        </p>
      </sec>
      <sec id="sec-4-5">
        <title>4.6. Classification Tasks</title>
        <p>Longer text can contain more information, therefore we often need more labels to classify
them and thus show two diferent classification types: Binary classification for two labels, and
multi-class classification for more than two class labels. At the end of the pipeline, we ensemble
the results of the summarization and classification tasks to get the final result.</p>
        <sec id="sec-4-5-1">
          <title>4.6.1. Binary Classification</title>
          <p>
            If the text is short, possibly containing a single sentence (as is often the case in social media),
the labels might be "true" and "false" or "toxic" and "non-toxic". However, other labels might be
used (such as "other"), turning the task into a multi-label classification. The chosen GermEval
datasets have two labels; thus, only the machine translation before is needed for fine-tuning.
We have decided for BERT [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ] as our English-based model, GBERT and GELECTRA [40] as
our German-based model, XLM-RoBERTa [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ] as our multilingual model with both German
and English input, and BERTweet [48] as our Twitter-based model. After five runs, they get
ensembled together in hard and soft majority voting (see Table 8). Then again, we choose the
best five model ensembles (in GermEval 2018 and 2019, the best three) and ensemble them in
three diferent ensembling strategies: Majority Voting (both hard and soft voting), Gradient
Boosting Machines and Logistic Regression (see Table 8).
          </p>
        </sec>
        <sec id="sec-4-5-2">
          <title>4.6.2. Multi-Class Classification</title>
          <p>
            Long texts typically contain more sentences and possibly a broader spread of topics. This
leads to classification tasks that go beyond a simple binary decision (e.g., one might consider
4https://huggingface.co/sshleifer/distilbart-cnn-12-6
"partially false" or "partially true"). The CheckThat! 2022 dataset has four diferent class labels
with imbalanced distributions. For the classification process, we use three large models: BERT
Uncased [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ], XLM-RoBERTa [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ], and T5-3B [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ].
          </p>
        </sec>
      </sec>
      <sec id="sec-4-6">
        <title>4.7. General Setup</title>
        <p>All of our experiments are conducted on the following datasets with the following GPUs:
the GermEval 2018 and 2019 datasets with GTX/RTX 1080/2080 Ti (11 GB VRAM) including
GermEval 2021 base models, Tesla V100S (32 GB VRAM) for the large models in the GermEval
2021 datasets, and the CheckThat! 2022 datasets with RTX A6000 with 48 GB VRAM. We use
the SimpleTransformers library5 for the T5 model and all other transformer models with the
Hugging Face Transformers library6. For the summarization task, we use the BERT Extractive
Summarizer library7 [42], and for machine translation, we use the deep-translator library8 in
combination with the free public Google Translate service9 and the pro version of the DeepL
Translator service10. Our hyperparameters are in Table 4.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>We observe that our approach is highly competitive and robust for both types of classification
and all datasets.</p>
      <sec id="sec-5-1">
        <title>5.1. Machine Translation</title>
        <p>We would first like to report some insights into the choice of Machine Translation tools. The
results show that for this experiment DeepL Translator appears to be a better choice than Google
Translate, but the score diference is very close, so both are solid choices (see Table 5). For the
GermEval datasets, we apply the machine translations of the DeepL Translator service. For the
CheckThat! 2022 dataset, since the maximum of the text, can be at 100,000 characters, we use
the free Google Translate service as a financial constraint. Since the service has an internal
character limit, we only take the first 5,000 characters for translation.</p>
        <p>5https://simpletransformers.ai/
6https://huggingface.co/transformers
7https://github.com/dmmiller612/bert-extractive-summarizer
8https://github.com/nidhalof/deep-translator
9https://translate.google.com/
10https://www.deepl.com/en/pro#developer</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Hyperparameter Optimization</title>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Splitting Methods</title>
        <p>As shown in Table 7, the diference after majority voting is minor, and thus both strategies are
eligible. If we look at each run, the diference is also very narrow. Thus, picking up a splitting
strategy is not essential and is not a deciding factor in the system architecture. We decide to
continue with random seeding.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Binary Classification and Ensembling</title>
        <p>For all GermEval datasets, we observe the potential for improvement over previously reported
SOTA results (see Table 8). For GermEval 2021 Subtask 1, the score improvement is noticeable
at 4.48% compared to the highest score reported so far. Except for GermEval 2021 Subtask 2,
where all results are more or less on par (which might in part be an issue with the gold standard
labels), all other results demonstrate the added value our approach ofers.</p>
        <p>Of all the ensembling strategies, the popular majority voting is still the most efective one.
Since Gradient Boosting Machines and Logistic Regression are both linear models, we expect
that the linear combination of their predictions will be more efective than the majority voting.
However, the results show that the majority voting is still the best strategy.</p>
      </sec>
      <sec id="sec-5-5">
        <title>5.5. Summarization and Multi-Class Classification</title>
        <p>As shown in Table 9, the combination of summarization and classification leads to noticeable
improvements (e.g., 5.63% for Task 3A). Unlike in the previous experiments, here we do not
apply ensembling, which could lead to further improvements in robustness and overall results.
An interesting observation here is the discrepancy between development sets and the test set
results.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Limitations and Future Work</title>
      <p>The results show that our approach is robust and achieves state-of-the-art performance on these
datasets. That ofers plenty of directions for future work. However, before we can start with
future work, we need to discuss the limitations of our approach.</p>
      <p>The first limitation is the fact that the hyperparameter search was random. A fixed scope of
hyperparameters might have led to better results for training. Another limitation is that we
have summarized every data point in the dataset. That means that even short text snippets
were summarized. We do not use the DeepL Translator service for all experiments because
DistilBART-CNN-12-6
(extractive)</p>
      <p>T5-3B
(abstractive)</p>
      <p>BERTlarge
XLM-Rlarge</p>
      <p>T5-3B
BERTlarge
XLM-Rlarge</p>
      <p>T5-3B</p>
      <p>SOTA
the free version is limited to 500,000 characters per month11 and one data point of the CLEF
CheckThat! 2022 dataset already hits 100,000 characters. Since the performance diference is
very close, we decided to use the free Google Translate service for the CheckThat! 2022 dataset.
We also want to warn about possible outputs caused by "model hallucination," which is not yet
usable for production.</p>
      <p>As future work, the investigation of text generation with bigger models like GPT-3 [54],
ChatGPT [55], PaLM [56], Flan-T5 [33], and others is interesting to see if our approach will
improve by simply having more parameters and more pre-trained data. Text generation tasks
like machine translation or summarization would benefit the increased accuracy of the models
and thus would lead to a real-world production environment to tackle fake news and hate
speech. Especially in the summarization task, we want to understand if summarizing text
snippets below 512 tokens makes a diference in performance. The increased performance by
summarization opens the question of why exactly it works and remains contentious. Another
open question is what the optimal amount of models for the ensemble is, where a correlation
between amount of dataset and diversity of models needs to be explored. Another important
question is how each module of the pipeline, especially summarization and machine translation,
work separately on a larger scale. Diferent benchmark datasets for each diferent tasks are
needed to investigate the performance of each module.</p>
      <p>11https://www.deepl.com/en/pro#developer</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>We propose a general architecture to deal with text classification in a cross-lingual context
tapping into resources available for high-resourced languages and making use of abstractive
and extractive summarization. We demonstrate the potential that this approach ofers using
existing non-English benchmark collections for fake news and hate speech classification. This
lays the groundwork for future work, which should look at a range of low-resource languages.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This work was supported by the project COURAGE: A Social Media Companion Safeguarding
and Educating Students funded by the Volkswagen Foundation, grant number 95564. We want
to thank all reviewers for their insightful comments that helped us to improve our work.
translation system: Bridging the gap between human and machine translation, ArXiv
abs/1609.08144 (2016).
[22] M. Junczys-Dowmunt, Microsoft translator at WMT 2019: Towards large-scale
documentlevel neural machine translation, in: Proceedings of the Fourth Conference on
Machine Translation (Volume 2: Shared Task Papers, Day 1), Association for Computational
Linguistics, Florence, Italy, 2019, pp. 225–233. URL: https://aclanthology.org/W19-5321.
doi:10.18653/v1/W19-5321.
[23] D. Coldewey, F. Lardinois, DeepL schools other online translators with
clever machine learning, 2017. URL: https://techcrunch.com/2017/08/29/
deepl-schools-other-online-translators-with-clever-machine-learning/.
[24] DeepL, Translation quality, ???? URL: https://www.deepl.com/en/quality.html.
[25] J. Fuchs, Spoofing Google Translate to Steal Credentials, 2022. URL: https://www.avanan.</p>
      <p>com/blog/spoofing-google-translate-to-steal-credentials.
[26] E. Montalbano, Cyberattackers spoof google translate in unique
phishing tactic, 2022. URL: https://www.darkreading.com/threat-intelligence/
cyberattackers-spoof-google-translate-unique-phishing-tactic.
[27] J. L. Neto, A. A. Freitas, C. A. A. Kaestner, Automatic text summarization using a machine
learning approach, in: Brazilian Symposium on Artificial Intelligence, 2002.
[28] R. Nallapati, F. Zhai, B. Zhou, Summarunner: A recurrent neural network based sequence
model for extractive summarization of documents, ArXiv abs/1611.04230 (2016).
[29] S. Chopra, M. Auli, A. M. Rush, Abstractive sentence summarization with attentive
recurrent neural networks, in: North American Chapter of the Association for Computational
Linguistics, 2016.
[30] R. Nallapati, B. Zhou, C. N. dos Santos, Çaglar Gülçehre, B. Xiang, Abstractive text
summarization using sequence-to-sequence rnns and beyond, in: Conference on Computational
Natural Language Learning, 2016.
[31] R. Paulus, C. Xiong, R. Socher, A deep reinforced model for abstractive summarization,</p>
      <p>ArXiv abs/1705.04304 (2017).
[32] P. Hartl, U. Kruschwitz, Applying automatic text summarization for fake news detection,
in: Proceedings of the 13th Language Resources and Evaluation Conference (LREC 2022),
2022, pp. 2702-–2713.
[33] H. W. Chung, L. Hou, S. Longpre, B. Zoph, Y. Tay, W. Fedus, E. Li, X. Wang, M. Dehghani,
S. Brahma, A. Webson, S. S. Gu, Z. Dai, M. Suzgun, X. Chen, A. Chowdhery, D. Valter,
S. Narang, G. Mishra, A. W. Yu, V. Zhao, Y. Huang, A. M. Dai, H. Yu, S. Petrov, E. H. hsin
Chi, J. Dean, J. Devlin, A. Roberts, D. Zhou, Q. Le, J. Wei, Scaling instruction-finetuned
language models, ArXiv abs/2210.11416 (2022).
[34] J. Opitz, S. Burst, Macro f1 and macro f1, arXiv preprint arXiv:1911.03347 (2019).
[35] M. Wiegand, M. Siegel, J. Ruppenhofer, Overview of the germeval 2018 shared task on the
identification of ofensive language, Proceedings of GermEval 2018, 14th Conference on
Natural Language Processing (KONVENS 2018), Vienna, Austria – September 21, 2018,
Austrian Academy of Sciences, Vienna, Austria, 2019, pp. 1 – 10. URL: https://nbn-resolving.
org/urn:nbn:de:bsz:mh39-84935.
[36] J. M. Struß, M. Siegel, J. Ruppenhofer, M. Wiegand, M. Klenner, Overview of germeval task
2, 2019 shared task on the identification of ofensive language, Preliminary proceedings
of the 15th Conference on Natural Language Processing (KONVENS 2019), October 9
– 11, 2019 at Friedrich-Alexander-Universität Erlangen-Nürnberg, German Society for
Computational Linguistics &amp; Language Technology und Friedrich-Alexander-Universität
Erlangen-Nürnberg, München [u.a.], 2019, pp. 352 – 363. URL: https://nbn-resolving.org/
urn:nbn:de:bsz:mh39-93197.
[37] J. Risch, A. Stoll, L. Wilms, M. Wiegand, Overview of the GermEval 2021 shared task on
the identification of toxic, engaging, and fact-claiming comments, in: Proceedings of the
GermEval 2021 Shared Task on the Identification of Toxic, Engaging, and Fact-Claiming
Comments, Association for Computational Linguistics, Duesseldorf, Germany, 2021, pp.
1–12.
[38] J. Köhler, G. K. Shahi, J. M. Struß, M. Wiegand, M. Siegel, T. Mandl, M. Schütz, Overview of
the clef-2022 checkthat! lab task 3 on fake news detection, Working Notes of CLEF (2022).
[39] C. Rafel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, P. J. Liu,
Exploring the limits of transfer learning with a unified text-to-text transformer, Journal of
Machine Learning Research 21 (2020) 1–67.
[40] B. Chan, S. Schweter, T. Möller, German’s next language model, in: Proceedings of the
28th International Conference on Computational Linguistics, International Committee on
Computational Linguistics, Barcelona, Spain (Online), 2020, pp. 6788–6796.
[41] T. Akiba, S. Sano, T. Yanase, T. Ohta, M. Koyama, Optuna: A next-generation
hyperparameter optimization framework, in: Proceedings of the 25th ACM SIGKDD International
Conference on Knowledge Discovery &amp; Data Mining, KDD ’19, Association for Computing
Machinery, New York, NY, USA, 2019, p. 2623–2631.
[42] D. Miller, Leveraging bert for extractive text summarization on lectures, arXiv preprint
arXiv:1906.04165 (2019).
[43] J. MacQueen, et al., Some methods for classification and analysis of multivariate
observations, in: Proceedings of the fifth Berkeley symposium on mathematical statistics and
probability, volume 1, Oakland, CA, USA, 1967, pp. 281–297.
[44] T. M. Kodinariya, P. R. Makwana, Review on determining number of cluster in k-means
clustering, International Journal 1 (2013) 90–95.
[45] S. Shleifer, A. M. Rush, Pre-trained summarization distillation, ArXiv abs/2010.13002
(2020).
[46] K. M. Hermann, T. Kocisky, E. Grefenstette, L. Espeholt, W. Kay, M. Suleyman, P.
Blunsom, Teaching machines to read and comprehend, in: C. Cortes, N. Lawrence, D. Lee,
M. Sugiyama, R. Garnett (Eds.), Advances in Neural Information Processing Systems,
volume 28, Curran Associates, Inc., 2015. URL: https://proceedings.neurips.cc/paper/2015/
ifle/afdec7005cc9f14302cd0474fd0f3c96-Paper.pdf.
[47] P. Shaw, J. Uszkoreit, A. Vaswani, Self-attention with relative position representations, in:
Proceedings of the 2018 Conference of the North American Chapter of the Association
for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers),
Association for Computational Linguistics, New Orleans, Louisiana, 2018, pp. 464–468.</p>
      <p>URL: https://aclanthology.org/N18-2074. doi:10.18653/v1/N18-2074.
[48] D. Q. Nguyen, T. Vu, A. T. Nguyen, BERTweet: A pre-trained language model for English
Tweets, in: Q. Liu, D. Schlangen (Eds.), Proceedings of the 2020 Conference on Empirical
Methods in Natural Language Processing: System Demonstrations, EMNLP 2020 - Demos,
Online, November 16-20, 2020, Association for Computational Linguistics, 2020, pp. 9–14.
[49] A. Paraschiv, D.-C. Cercel, Upb at germeval-2019 task 2: Bert-based ofensive language
classification of german tweets., in: KONVENS, 2019.
[50] T. Bornheim, N. Grieger, S. Bialonski, FHAC at GermEval 2021: Identifying German toxic,
engaging, and fact-claiming comments with ensemble learning, in: Proceedings of the
GermEval 2021 Shared Task on the Identification of Toxic, Engaging, and Fact-Claiming
Comments, Association for Computational Linguistics, Duesseldorf, Germany, 2021, pp.
105–111.
[51] N. Hildebrandt, B. Boenninghof, D. Orth, C. Schymura, Data science kitchen at
GermEval 2021: A fine selection of hand-picked features, delivered fresh from the oven, in:
Proceedings of the GermEval 2021 Shared Task on the Identification of Toxic, Engaging,
and Fact-Claiming Comments, Association for Computational Linguistics, Duesseldorf,
Germany, 2021, pp. 88–94. URL: https://aclanthology.org/2021.germeval-1.13.
[52] B. Taboubi, M. A. B. Nessir, H. Haddad, icompass at checkthat! 2022: combining deep
language models for fake news detection, Working Notes of CLEF (2022).
[53] H. N. Tran, U. Kruschwitz, ur-iw-hnt at CheckThat! 2022: cross-lingual text summarization
for fake news detection, in: G. Faggioli, N. Ferro, A. Hanbury, M. Potthast (Eds.), Working
Notes of CLEF 2022 - Conference and Labs of the Evaluation Forum, CLEF ’2022, Bologna,
Italy, 2022.
[54] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan,
P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan,
R. Child, A. Ramesh, D. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin,
S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, D. Amodei,
Language models are few-shot learners, in: H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan,
H. Lin (Eds.), Advances in Neural Information Processing Systems, volume 33, Curran
Associates, Inc., 2020, pp. 1877–1901. URL: https://proceedings.neurips.cc/paper/2020/file/
1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf.
[55] OpenAI, ChatGPT: Optimizing Language Models for Dialogue, 2022. URL: https://openai.</p>
      <p>com/blog/chatgpt/.
[56] A. Chowdhery, S. Narang, J. Devlin, M. Bosma, G. Mishra, A. Roberts, P. Barham, H. W.</p>
      <p>Chung, C. Sutton, S. Gehrmann, P. Schuh, K. Shi, S. Tsvyashchenko, J. Maynez, A. Rao,
P. Barnes, Y. Tay, N. Shazeer, V. Prabhakaran, E. Reif, N. Du, B. Hutchinson, R. Pope,
J. Bradbury, J. Austin, M. Isard, G. Gur-Ari, P. Yin, T. Duke, A. Levskaya, S. Ghemawat,
S. Dev, H. Michalewski, X. Garcia, V. Misra, K. Robinson, L. Fedus, D. Zhou, D. Ippolito,
D. Luan, H. Lim, B. Zoph, A. Spiridonov, R. Sepassi, D. Dohan, S. Agrawal, M. Omernick,
A. M. Dai, T. S. Pillai, M. Pellat, A. Lewkowycz, E. Moreira, R. Child, O. Polozov, K. Lee,
Z. Zhou, X. Wang, B. Saeta, M. Diaz, O. Firat, M. Catasta, J. Wei, K. Meier-Hellstern,
D. Eck, J. Dean, S. Petrov, N. Fiedel, PaLM: Scaling Language Modeling with Pathways,
arxiv:2204.02311 (2022). URL: https://arxiv.org/abs/2204.02311.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. P. A.</given-names>
            <surname>Corney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barrón-Cedeño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Papotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D. S.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <article-title>Automated fact-checking for assisting human fact-checkers</article-title>
          , in: Z.
          <string-name>
            <surname>Zhou</surname>
          </string-name>
          (Ed.),
          <source>Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI</source>
          <year>2021</year>
          , Virtual Event / Montreal, Canada,
          <fpage>19</fpage>
          -27
          <source>August</source>
          <year>2021</year>
          ,
          <article-title>ijcai</article-title>
          .org,
          <year>2021</year>
          , pp.
          <fpage>4551</fpage>
          -
          <lpage>4558</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <article-title>Social media and attitude change: Information booming promote or resist persuasion?</article-title>
          ,
          <source>Frontiers in Psychology</source>
          <volume>12</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Fersko</surname>
          </string-name>
          ,
          <article-title>Is social media bad for teens' mental health</article-title>
          ?,
          <year>2018</year>
          . URL: https://www.unicef. org/stories/social
          <article-title>-media-bad-teens-mental-health.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Newman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fletcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Eddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Nielsen</surname>
          </string-name>
          ,
          <source>Digital news report</source>
          <year>2022</year>
          ,
          <year>2022</year>
          . URL: https://reutersinstitute.politics.ox.ac.uk/sites/default/files/2022-06/ Digital_News-Report_
          <year>2022</year>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <article-title>[5] Reuters Institute for the Study of Journalism, Share of adults who trust news media most of the time in selected countries worldwide as of February 2022</article-title>
          [Graph],
          <year>2022</year>
          . URL: https: //www.statista.com/statistics/308468/importance-brand
          <article-title>-journalist-creating-trust-news/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/ N19-1423. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>G. L. de Holanda Coelho</surname>
            ,
            <given-names>P. H. P.</given-names>
          </string-name>
          <string-name>
            <surname>Hanel</surname>
            ,
            <given-names>R. P.</given-names>
          </string-name>
          <string-name>
            <surname>Monteiro</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Vilar</surname>
            ,
            <given-names>V. V.</given-names>
          </string-name>
          <string-name>
            <surname>Gouveia</surname>
          </string-name>
          ,
          <article-title>The dark side of human values: How values are related to bright and dark personality traits</article-title>
          ,
          <source>The Spanish Journal of Psychology</source>
          <volume>24</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Qian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ruchansky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Liu,
          <article-title>Combating fake news: A survey on identification and mitigation techniques</article-title>
          ,
          <source>ACM Transactions on Intelligent Systems and Technology (TIST) 10</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>42</lpage>
          . Publisher: ACM New York, NY, USA.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          , NIPS'17, Curran Associates Inc., USA,
          <year>2017</year>
          , pp.
          <fpage>6000</fpage>
          -
          <lpage>6010</lpage>
          . URL: http://dl.acm.org/citation.cfm?id=
          <volume>3295222</volume>
          .
          <fpage>3295349</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, BART:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>67</lpage>
          . URL: http://jmlr.org/papers/v21/
          <fpage>20</fpage>
          -
          <lpage>074</lpage>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Basile</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mitrović</surname>
          </string-name>
          , M. Granitzer,
          <article-title>HateBERT: Retraining BERT for abusive language detection in English</article-title>
          ,
          <source>in: Proceedings of the 5th Workshop on Online Abuse and Harms (WOAH</source>
          <year>2021</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>17</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Waseem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hovy</surname>
          </string-name>
          ,
          <article-title>Hateful symbols or hateful people? predictive features for hate speech detection on Twitter</article-title>
          ,
          <source>in: Proceedings of the NAACL Student Research Workshop</source>
          , Association for Computational Linguistics, San Diego, California,
          <year>2016</year>
          , pp.
          <fpage>88</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>T.</given-names>
            <surname>Murayama</surname>
          </string-name>
          ,
          <article-title>Dataset of fake news detection and fact verification: A survey</article-title>
          ,
          <source>ArXiv abs/2111</source>
          .03299 (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Arnold</surname>
          </string-name>
          ,
          <article-title>The challenges of online fact checking</article-title>
          ,
          <source>Technical Report</source>
          , Full Fact, London, UK,
          <year>2020</year>
          . URL: https://fullfact.org/media/uploads/coof-2020.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Unsupervised cross-lingual representation learning at scale, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>8440</fpage>
          -
          <lpage>8451</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>747</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>747</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>G.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <article-title>A study of multilingual toxic text detection approaches under imbalanced sample distribution</article-title>
          ,
          <source>Information (Switzerland) 12</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18] CSA Research,
          <source>Survey of 8,709 Consumers in 29 Countries Finds that 76% Prefer Purchasing Products with Information in their Own Language</source>
          ,
          <year>2020</year>
          . URL: https://csa-research.com/Blogs-Events/
          <article-title>CSA-in-the-</article-title>
          <string-name>
            <surname>Media/</surname>
          </string-name>
          Press-Releases/
          <article-title>Consumers-Prefer-their-</article-title>
          <string-name>
            <surname>Own-Language</surname>
          </string-name>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <article-title>Global Market Insights Inc., Machine translation market size worldwide, from 2016 to 2024 (in million U.S. dollars</article-title>
          ) [Graph],
          <year>2017</year>
          . URL: https://www.statista.com/statistics/748358/ worldwide-machine
          <article-title>-translation-market-size/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wadhwani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gankar</surname>
          </string-name>
          ,
          <article-title>Machine translation market size: Industry analysis,</article-title>
          <year>2022</year>
          -
          <fpage>2030</fpage>
          ,
          <year>2022</year>
          . URL: https://www.gminsights.com/industry
          <article-title>-analysis/ machine-translation-market-size.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Norouzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Macherey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krikun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Macherey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Klingner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , X. Liu,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gouws</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kudo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kazawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Stevens</surname>
          </string-name>
          , G. Kurian,
          <string-name>
            <given-names>N.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Riesa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rudnick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. S.</given-names>
            <surname>Corrado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hughes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Google's neural machine</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>