<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Conference and Labs of the Evaluation Forum, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Enhancing Writing Style Change Detection using Transformer-based Models and Data Augmentation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ahmad Hashemi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wei Shi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Information Technology, Carleton University</institution>
          ,
          <addr-line>Ottawa</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>1</volume>
      <fpage>8</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>This paper presents our participation in the Style Change Detection task for PAN at CLEF 2023. The primary goal of this task is to identify alterations in writing style at the paragraph level within a provided document that has been authored by multiple writers. The task comprises three sub-tasks that difer in dificulty levels, primarily based on the diversity of topics addressed within the paragraphs. To address these sub-tasks, we investigate the efectiveness of fine-tuning diferent pre-trained transformer-based models, with a particular emphasis on RoBERTa. Additionally, we employ data augmentation techniques to enhance the performance of our models. Furthermore, we incorporate ensemble modeling to further improve the accuracy and robustness of our style change detection system. In the competition, our provided models achieved the first rank in terms of F1 score for two of the sub-tasks, and secured the second position for the remaining sub-task.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Authorship Attribution</kwd>
        <kwd>Plagiarism Detection</kwd>
        <kwd>Ensemble Learning</kwd>
        <kwd>Transformers</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Multi-Author writing style analysis is an interesting area of study that focuses on analyzing
documents that have been written by multiple authors. It involves a range of tasks, such as
determining whether a document is the product of a single author or multiple authors [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], as well
as investigating the occurrence and positioning of style changes in multi-authored documents
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The primary motivation behind Multi-Author writing style analysis is that it enables the
identification of positions where authors switch within a text, allowing for the detection of
plagiarism and the verification of claimed authorship, even in the absence of comparison texts.
Style change detection also assists in uncovering gift authorships and can contribute to the
development of innovative technologies for writing support [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        The style change detection task introduced by PAN [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] this year focuses on detecting writing
style changes at the paragraph level in a given text document. The objective is to detect style
changes between consecutive paragraphs, assessing whether there was a transition in writing
style. The task provides datasets of three dificulty levels: easy (subtask 1), medium (subtask 2),
and hard (subtask 3. In the easy level, paragraphs cover a range of topics, allowing approaches to
utilize topic information for detecting authorship changes. The medium level features a smaller
topical variety, requiring the approach to focus more on style to efectively solve the detection
task. The hard level consists of paragraphs on the same topic throughout the document [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        With the advancements in natural language processing techniques, many researchers have
directed their attention toward various tasks within the realm of digital text forensics. These
tasks cover a wide range, including the detection of fake news [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], spam [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and hate speech
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], as well as author profiling [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], authorship attribution [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and style change detection
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Earlier studies primarily employed feature-based approaches, involving the extraction of
stylistic features followed by applying traditional machine learning or deep learning algorithms.
However, the emergence of pre-trained transformer-based models, with their remarkable
capabilities, has led to a shift in focus for many recent studies. These studies now mostly center
around fine-tuning pre-trained models to customize them for their respective tasks.
      </p>
      <p>In the domain of style change detection, early attempts primarily revolved around the
extraction of stylometric features. For instance, Eissen and Stein [12] employed word frequency
classes to diferentiate between distinct writing styles in order to investigate intrinsic plagiarism
detection. Bensalem et al. [13] utilized n-grams to identify authorial style changes, while Gianella
[14] applied Bayesian modeling techniques to segment a document based on authorship. Another
approach [15], involved the use of neural networks in conjunction with various stylometric
features.</p>
      <p>More recent approaches in style change detection have mostly leveraged pre-trained models,
although some still incorporate stylometric features. For example, Iyer and Vosoughi [16]
employed Google AI’s BERT pre-trained bidirectional models to tokenize and generate sentence
embeddings, which were then utilized to train a random forest classifier for the PAN’s SCD
task. For the PAN SCD task of 2021, Singh et al. [17] extracted stylometric features from each
paragraph and used the absolute diferences between the feature vectors to train a Logistic
Regression classifier to determine if two paragraphs were written by the same author. Lin et
al. [18] fine-tuned transformer models such as BERT, RoBERTa, and ALBERT, along with their
classifiers, to measure the similarity between paragraphs or sentences for authorship analysis
in the most recent PAN style change detection task in 2022.</p>
      <p>In our work, we investigate the efectiveness of leveraging three pre-trained
transformerbased models in detecting style changes between consecutive paragraphs, where both the author
and topic change, as well as cases where only the author changes while the topic remains the
same. Additionally, we apply data augmentation techniques and employ ensemble modeling to
enhance the performance of our approach.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>For each of the subtasks, a separate dataset has been provided. These datasets comprise
multiple documents, with each document containing some paragraphs. For every document, a
corresponding ground truth file is available, providing two pieces of information: 1) the number
Dataset
Training 1
Training 2
Training 3
Validation 1
Validation 2
Validation 3
of authors associated with the document, and 2) the identification of consecutive paragraphs
where a style change has occurred, indicating a transition in authorship. Each subtask’s dataset
has been divided into a training set and a validation set.</p>
      <p>For our experimental setup, we treated every pair of consecutive paragraphs as a sample
and concatenated them. Each sample was assigned a label indicating whether a style change
occurred between the two paragraphs (labeled as 1) or if they were written by the same author
(labeled as 0). Further details and statistics about the datasets in our experimental setup can be
found in Table 1.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <sec id="sec-4-1">
        <title>4.1. Data preparation</title>
        <p>To create the samples, we begin by concatenating two consecutive paragraphs within each
document using a separator token. Next, we assign the associated binary label indicating
whether a style change occurs between the two paragraphs. This allows us to transform
the task into a binary classification problem. To prepare the samples for fine-tuning the
pretrained transformer-based models (specifically BERT, RoBERTa, and ELECTRA), we employ the
corresponding tokenizer associated with each model. However, it is important to note that these
models have limitations regarding the maximum input sequence size, typically set at 512 tokens.
To address this limitation, we analyze the datasets and find that in each dataset only a few
samples exceed the maximum token limit. Therefore, we opt to truncate the longer samples. To
ensure equal attention is given to each paragraph in the sample, we adopt a truncation strategy
that involves removing tokens from both ends of the sequence.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Pre-trained Transformer Models</title>
        <p>Pre-trained transformer models are powerful language models that are trained on massive
amounts of text data. They learn to understand the structure and patterns of language, enabling
them to generate high-quality text and perform various natural language processing (NLP)
tasks. Fine-tuning the pre-trained models allows us to leverage their language understanding
capabilities and transfer the knowledge they have acquired from their extensive pre-training
to our specific task. In our study, we employed three popular pre-trained transformer models,
namely BERT, RoBERTa, and ELECTRA.</p>
        <p>BERT (Bidirectional Encoder Representations from Transformers) is a revolutionary model
that introduced the concept of bidirectional context to capture the dependencies between words.
It utilizes a transformer architecture and pre-training tasks such as masked language modeling
to learn contextualized representations of words [19]. RoBERTa, an extension of BERT, further
enhanced the pre-training process by utilizing additional training data and applying dynamic
masking strategies. This enabled RoBERTa to achieve even better performance across a range of
NLP tasks [20]. ELECTRA (Eficiently Learning an Encoder that Classifies Token Replacements
Accurately) introduced a novel pre-training method called "discriminative masked language
modeling." It improves eficiency by generating synthetic training data and training the model
to diferentiate between original and replaced tokens [21].</p>
        <p>To adapt these pre-trained models for our specific task of style change detection, we added a
binary classification layer on top of each model. This allowed the models to learn and classify
whether a style change occurred between the consecutive paragraphs in each sample. For
each subtask, we fine-tuned the models using the associated dataset to consider the unique
characteristics of each subtask.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Data augmentation</title>
        <p>To explore the potential impact of generating more samples while disregarding the
consecutiveness and order of the paragraphs, we conduct an investigation. We transform the task into
determining whether the same author wrote two concatenated paragraphs (not necessarily
consecutive or in order). This allows us to explore diferent approaches for data augmentation.
In our first approach, we swap the order of the two paragraphs in each sample, creating a new
sample with the same label to double the number of samples. Additionally, we leverage the
metadata provided in the datasets to identify paragraphs written by the same authors that are
not necessarily consecutive. Using the metadata, which includes the number of authors in each
document, we employ the Algorithm 1 for each document to retrieve such pairs of paragraphs
and augment the data.</p>
        <p>As Algorithm 1 demonstrates, we incorporate additional non-consecutive pairs of paragraphs
into our sample set and assign them labels based on the inferred relationships. For example, if
there are three consecutive paragraphs without a style change, we can infer that the first and
third paragraphs are written by the same author. Similarly, if there are style changes between
the first and second paragraphs and between the second and third paragraphs, we can deduce
that the authors of the first and third paragraphs are diferent, given that the number of authors
in the document exceeds the number of style changes by one.</p>
        <p>After data augmentation, we utilize our extended sample sets for each subtask to fine-tune
the pre-trained model following the same process explained earlier. It is important to note
that this augmentation introduces a change in the nature of the train set, deviating from the
consecutive and ordered structure that exists in the test data. However, the advantage of having
a larger dataset can improve the model’s ability to capture patterns across paragraphs written
by the same author.</p>
        <p>end
 ←  + 1 ;
end
 ←  + 1 ;
Algorithm 1: Pseudocode for data augmentation based on non-consecutive paragraph
pairs
if style_changes.count(1) = (authors_count − 1) then
for  in range(len(paragraphs) − 1) do
 ←  + 1 ; // set the next paragraph index
while  &lt; len(paragraphs) and style_changes[ − 1] = 0 do
; // while same author
if  &gt;  + 1 then
samples.append(paragraphs[] + seperator + paragraphs[])
labels.append(0) ; // same author</p>
        <p>// move to the next paragraph
end
while  &lt; len(paragraphs) do
if  &gt;  + 1 then
samples.append(paragraphs[] + seperator + paragraphs[])
labels.append(1) ; // style change</p>
        <p>// move to the next paragraph
end
end
end</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Generalization and Ensemble modeling</title>
        <p>Given the similarity in the nature of datasets across diferent subtasks, particularly in terms
of the presence of style changes, we explore the potential benefits of leveraging datasets from
other subtasks to enhance the model performance for a specific subtask. For example, Task
2 and Task 3 share similarities as they both involve style changes occurring while the topic
remains consistent. Similarly, Task 2 and Task 1 exhibit similarities as they both encompass
style changes alongside topic transitions. We believe that the datasets from Task 1 and Task
3 can provide valuable insights to Task 2, as they encompass scenarios that align with this
task. Accordingly, our investigation involves not only fine-tuning the model using the provided
dataset specific to the subtask but also incorporating additional samples from other subtasks’
datasets to assess the impact of having a generalized model.</p>
        <p>To maximize the potential benefits ofered by the various approaches mentioned, we employ
an ensemble strategy based on the majority voting approach. For each subtask, we develop
three models, all based on fine-tuning the RoBERTa pre-trained model, which we will discuss
in subsequent sections as the most efective among the investigated pre-trained models for all
subtasks. The first model utilizes the initial samples exclusively from the task-specific dataset for
ifne-tuning. The second model incorporates augmented samples derived from the task-specific
dataset, thereby expanding the training data. Lastly, the third model combines all the original
samples from the other datasets in addition to the task-specific dataset to provide a generalized
model. By ensembling these models, we aim to leverage the strengths of each approach and
enhance the overall performance of our style change detection system.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments</title>
      <sec id="sec-5-1">
        <title>5.1. Experimental settings</title>
        <p>We downloaded the large versions of pre-trained BERT, RoBERTa, and ELECTRA models from
HuggingFace [22]. The implementation and fine-tuning of these models were conducted on a
server equipped with an NVIDIA A100 GPU. To optimize the performance of our models, we
selected hyperparameter values as follows: We set the maximum sequence length to 512, the
learning rate to 0.00001, the batch size to 16, and the number of epochs to 10.</p>
        <p>To assess the efectiveness of the models for each subtask, we evaluate their performance by
computing the F1 score on the provided evaluation set. The F1 score is calculated based on the
predictions made by the models for detecting style changes between consecutive paragraph
pairs within the evaluation set of each subtask. After conducting our experiments and obtaining
results on the evaluation sets, we select the best-performing model for each subtask and run
the model on an unseen test set using the TIRA platform [23].</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Results</title>
        <p>The results of our experiments are presented in this section. Firstly, Table 2 displays the
performance comparison of fine-tuning pre-trained models on the original task-specific datasets
for each subtask. The findings indicate that RoBERTa consistently outperforms both BERT and
ELECTRA across all subtasks, establishing its superiority in the task of style change detection.
As a result, we have selected RoBERTa as the preferred approach for our style change detection
system.</p>
        <p>Moving on, Table 3 presents the performance evaluation results of our RoBERTa-based
experiments on the evaluation set. It highlights the F1 scores for each subtask and approach.
The findings reveal that fine-tuning RoBERTa on the original task-specific dataset yields the
highest F1 scores among all the provided approaches for subtasks 1 and 3. As can be seen, the
performance of the generalized model trained on all the datasets drops for subtask 1 and subtask
3, which aligns with our expectations as these subtasks possess exclusive characteristics that
may not benefit from additional data from the other subtasks. On the other hand, subtask 2
exhibits similarities to both subtask 1 (style changes along with topic changes) and subtask
3 (style changes without topic change), which explains why the performance drop for the
generalized model is less significant in subtask 2.</p>
        <p>Furthermore, incorporating augmented data for subtask 3 leads to a notable performance
drop, suggesting that considering paragraph order and consecutiveness is not negligible for
detecting style changes when the topic is consistent. However, the utilization of augmented data
produces competitive performance compared to only using the original samples for subtasks 1
and 2. Notably, the ensemble model for subtask 2 achieves the best F1 score, indicating that the
individual models capture complementary patterns that contribute to the overall performance
improvement.</p>
        <p>Based on the validation set results, we select the ensemble approach for subtask 2 and the
basic RoBERTa models for subtasks 1 and 3 to perform on the final test sets. The final results
of our selected models on the unseen test sets are presented in Table 4. According to the
provided results reported by the competition organizers, our provided models for subtasks 1
and 2 outperformed all other participants’ models, securing the first position. For subtask 3, we
achieved the second position.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In conclusion, our investigation aimed to enhance style change detection in textual documents
through various techniques and approaches. We explored data augmentation strategies to
generate non-consecutive paragraph pairs, allowing the model to learn patterns beyond
sequential and ordered paragraphs. We found that fine-tuned RoBERTa models outperformed BERT
and ELECTRA in all the task dificulty levels, demonstrating their efectiveness in capturing
style changes. The results also highlighted the importance of considering the unique nature of
each task, as incorporating additional data from unrelated tasks did not necessarily improve
performance.</p>
      <p>Furthermore, the ensemble approach proved to be valuable in capturing complementary
patterns, particularly for subtask 2, where the nature of the subtask contained similarities to
both the other subtasks. This ensemble model outperformed individual models,
emphasizing the benefits of combining multiple perspectives. We believe these insights contribute to
advancements in authorship attribution and plagiarism detection applications.
R. Ortega-Bueno, P. Pęzik, M. Potthast, et al., Overview of pan 2022: Authorship
verification, profiling irony and stereotype spreaders, and style change detection, in: Experimental
IR Meets Multilinguality, Multimodality, and Interaction: 13th International Conference
of the CLEF Association, CLEF 2022, Bologna, Italy, September 5–8, 2022, Proceedings,
Springer, 2022, pp. 382–394.
[12] S. M. z. Eissen, B. Stein, Intrinsic plagiarism detection, in: Advances in Information
Retrieval: 28th European Conference on IR Research, ECIR 2006, London, UK, April 10-12,
2006. Proceedings 28, Springer, 2006, pp. 565–569.
[13] I. Bensalem, P. Rosso, S. Chikhi, Intrinsic plagiarism detection using n-gram classes, in:
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing
(EMNLP), 2014, pp. 1459–1464.
[14] C. Giannella, An improved algorithm for unsupervised decomposition of a multi-author
document, Journal of the Association for Information Science and Technology 67 (2016)
400–411.
[15] N. Graham, G. Hirst, B. Marthi, Segmenting documents by stylistic character, Natural</p>
      <p>Language Engineering 11 (2005) 397–415.
[16] A. Iyer, S. Vosoughi, Style change detection using bert., CLEF (Working Notes) 93 (2020)
106.
[17] R. Singh, J. Weerasinghe, R. Greenstadt, Writing style change detection on multi-author
documents., in: CLEF (Working Notes), 2021, pp. 2137–2145.
[18] T.-M. Lin, C.-Y. Chen, Y.-W. Tzeng, L.-H. Lee, Ensemble pre-trained transformer models
for writing style change detection, CLEF, 2022.
[19] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, arXiv preprint arXiv:1810.04805 (2018).
[20] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer,
V. Stoyanov, Roberta: A robustly optimized bert pretraining approach, arXiv preprint
arXiv:1907.11692 (2019).
[21] K. Clark, M.-T. Luong, Q. V. Le, C. D. Manning, Electra: Pre-training text encoders as
discriminators rather than generators, arXiv preprint arXiv:2003.10555 (2020).
[22] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf,
M. Funtowicz, et al., Huggingface’s transformers: State-of-the-art natural language
processing, arXiv preprint arXiv:1910.03771 (2019).
[23] M. Fröbe, M. Wiegmann, N. Kolyada, B. Grahm, T. Elstner, F. Loebe, M. Hagen, B. Stein,
M. Potthast, Continuous Integration for Reproducible Shared Tasks with TIRA.io, in:
J. Kamps, L. Goeuriot, F. Crestani, M. Maistro, H. Joho, B. Davis, C. Gurrin, U. Kruschwitz,
A. Caputo (Eds.), Advances in Information Retrieval. 45th European Conference on IR
Research (ECIR 2023), Lecture Notes in Computer Science, Springer, Berlin Heidelberg
New York, 2023, pp. 236–241.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kestemont</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tschuggnall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Daelemans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Potthast, Overview of the author identification task at pan-2018: cross-domain authorship attribution and style change detection</article-title>
          ,
          <source>in: Working Notes Papers of the CLEF</source>
          <year>2018</year>
          <article-title>Evaluation Labs</article-title>
          . Avignon, France,
          <source>September 10-14</source>
          ,
          <year>2018</year>
          /Cappellato, Linda [edit.]; et al.,
          <year>2018</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          , G. Specht,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the style change detection task at pan 2020</article-title>
          .,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          (Working Notes)
          <volume>93</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[3] Style Change Detection Task at CLEF</source>
          <year>2023</year>
          , https://pan.webis.de/clef23/pan23-web/ style-change-detection.html,
          <year>2023</year>
          .
          <source>Accessed: July 2</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Borrego-Obrador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chinea-Ríos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Franco-Salvador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Heini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kredens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pęzik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wolska</surname>
          </string-name>
          , , E. Zangerle, Overview of PAN 2023:
          <article-title>Authorship Verification, Multi-Author Writing Style Analysis, Profiling Cryptocurrency Influencers, and Trigger Detection</article-title>
          , in: A.
          <string-name>
            <surname>Arampatzis</surname>
            , E. Kanoulas,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Tsikrika</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          <string-name>
            <surname>Stefanos Vrochidis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Aliannejadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Vlachos</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the Multi-Author Writing Style Analysis Task at PAN 2023</article-title>
          , in: M.
          <string-name>
            <surname>Aliannejadi</surname>
            , G. Faggioli,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          , M. Vlachos (Eds.), Working Notes of CLEF 2023 -
          <article-title>Conference and Labs of the Evaluation Forum, CEUR-</article-title>
          <string-name>
            <surname>WS</surname>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zafarani</surname>
          </string-name>
          ,
          <article-title>A survey of fake news: Fundamental theories, detection methods, and opportunities</article-title>
          ,
          <source>ACM Computing Surveys (CSUR) 53</source>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>40</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Bhatia</surname>
          </string-name>
          ,
          <article-title>A review on social spam detection: Challenges, open issues, and future directions</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>186</volume>
          (
          <year>2021</year>
          )
          <fpage>115742</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zubiaga</surname>
          </string-name>
          ,
          <article-title>Towards generalisable hate speech detection: a review on obstacles and solutions</article-title>
          ,
          <source>PeerJ Computer Science</source>
          <volume>7</volume>
          (
          <year>2021</year>
          )
          <article-title>e598</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>HaCohen-Kerner</surname>
          </string-name>
          ,
          <article-title>Survey on profiling age and gender of text authors</article-title>
          ,
          <source>Expert Systems with Applications</source>
          (
          <year>2022</year>
          )
          <fpage>117140</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <article-title>A review on authorship attribution in text mining</article-title>
          ,
          <source>Wiley Interdisciplinary Reviews: Computational Statistics</source>
          <volume>15</volume>
          (
          <year>2023</year>
          )
          <article-title>e1584</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Fersini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Heini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kestemont</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kredens</surname>
          </string-name>
          , M. Mayerl,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>