<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the Multi-Author Writing Style Analysis Task at PAN 2024</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Eva Zangerle</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maximilian Mayerl</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martin Potthast</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benno Stein</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bauhaus-Universität Weimar</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Applied Sciences BFI Vienna</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Innsbruck</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Kassel</institution>
          ,
          <addr-line>hessian.AI, and ScaDS.AI</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Analyzing the writing style of individual authors in texts in which several authors are involved is a fundamental task in attributing authorship and detecting plagiarism, as it makes it possible to identify the points at which authorship changes. This year's multi-author writing style analysis task focuses on identifying all instances of paragraph-level writing style changes within a given text. We provide datasets with three diferent degrees of topical homogeneity to investigate how diferent degrees of topic consistency afect the detection of writing style changes. This paper gives an overview of the task, its definition and the data used, the approaches proposed by the participants, and the results obtained.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Writing style analysis requires an intrinsic analysis of author writing styles: no information on
authorship from external sources is used. The core of intrinsic writing style analysis is the computation of
stylistic profiles on the basis of text features. By computing similarities between the profiles of text
segments, changes in writing style can be detected, which is an indicator for a potential change in
authorship [1, 2]. Profiles are based on features that describe the writing style of authors, including
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) lexical features (character n-grams (e.g., [3, 4, 5]), word frequencies (e.g., [6]), and average word or
sentence lengths (e.g., [7])), (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) syntactic features (such as part-of-speech tag frequencies and structures
(e.g., [8]), or grammar trees (e.g., [9])), or (3) structural features (e.g., indentation usage (e.g., [7])).
These profiles are then used to match text fragments written by the same author [ 10], cluster authorial
threads [11, 12, 13, 14], or to predict the number of authors [15].
      </p>
      <p>The multi-author writing style analysis task, formerly known as the style change detection task, has
been organized at PAN since 2016. Over the years, the tasks and the data used have constantly evolved.
However, the main objective has remained the same: analyzing authors’ writing styles to identify the
positions at which authorship changes in texts by multiple authors. Since the first edition in 2016, we
have seen significant progress in the results.</p>
      <p>In the first editions of the PAN task in 2016, participants were asked to identify and cluster text
segments by author [16]. In 2017, the aim was to recognize whether a document was written by several
authors [17]; if there were several authors, the participants were asked to indicate the exact positions
of these changes. In 2018, the task was to distinguish between documents from single authors and
documents from multiple authors [18]. In 2019, the task was extended to also predict the number of
authors [19]. Since 2020, style changes had to be identified at the paragraph level [ 20, 21], and in 2021
also the authors had to be assigned to paragraphs [21]. In 2022, the task was extended to detect changes
not only at the paragraph level, but even at the sentence level [22], while in 2023 the recognition was
performed at the paragraph level again [23].</p>
      <p>In recent years, large language models (LLMs) have made considerable progress; they are inherently
well suited to analyzing writing styles with multiple authors. For example, while in 2018 the winning
approach was based on the extraction of lexical and syntactic features [24] and a stacking ensemble
classifier, from 2020 the majority of submitted approaches are based on LLMs fine-tuned on the training
data [25, 26, 27, 28].</p>
      <p>For the 2024 edition of the writing style analysis task at PAN, we ask participants to detec any changes
in writing style at the paragraph level. We provide three datasets with increasing topical homogeneity
of the paragraphs and thus increasing dificulty.</p>
      <p>The remainder of this paper is structured as follows. Section 2 presents the PAN 2024 multi-author
writing style analysis task, the data used, and the evaluation setup. Section 3 surveys the participants’
submissions, while Section 4 presents an analysis and comparison of the achieved results, and Section 5
concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Style Change Detection Task</title>
      <sec id="sec-2-1">
        <title>2.1. Task Definition</title>
        <p>Participants of this year’s multi-author writing style analysis task were asked to solve the following
intrinsic style change detection task: For a given text, find all positions of writing style change at
the paragraph level, i.e., for each pair of consecutive paragraphs, assess whether there was a style
change. We control the dificulty of the task by managing the variety of topics in the given documents.
Participants are provided with data sets with three levels of dificulty:</p>
        <p>easy
medium</p>
        <p>The document covers a range of topics, allowing topical changes between paragraphs to be
used as style change signals.</p>
        <p>The document exhibits minimal topical variety (though some still exists), requiring the
approaches to focus on stylistic features for the task.
hard</p>
        <p>The paragraphs of a document all are on the same topic.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Dataset</title>
        <p>Continuing our eforts from the 2023 competition, this year’s data set for the multi-author writing style
analysis task is again based on user posts on Reddit, a popular social messaging platform.</p>
        <p>For the generation of the dataset, we selected a set of subreddits (topical sub-threads on Reddit)
that we expected to yield longer and more detailed texts by individual users: r/worldnews, r/politics,
r/askhistorians, and r/legaladvice. After scraping these threads, we applied cleaning and preprocessing
steps to the gathered texts. This included removing citations, markdown, emojis, hyperlinks, multiple
line breaks, and extra whitespace.</p>
        <p>
          The texts were divided into individual paragraphs. Paragraphs originating from the same Reddit
thread were combined into documents for the datasets, ensuring minimal topical coherence within each
document. Style changes were introduced by randomly selecting paragraphs from diferent authors
within the thread. To control for topical variability and thus the extent to which thematic aspects can
be used as a style change signal (and thus the complexity of the task), we consider the semantic and
stylistic properties of the paragraphs. The paragraphs are arranged based on these pair-wise paragraph
similarities, configuring these similarities to be (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) “large” for the easy dataset, (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) “moderate” for the
medium dataset, and (3) “small” for the hard dataset.
        </p>
        <p>We configured the dataset creation process to create documents written by two to four authors
to ensure an even distribution of documents according to the number of authors. Each of the three
resulting datasets contains 6,000 documents, each split into a training dataset (70% of all documents), a
validation dataset (15% of all documents), and a test dataset (15% of all documents), which is held back
until the evaluation phase of the task.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Performance Measures</title>
        <p>We evaluate the submitted approaches independently for each of the three datasets. Each approach is
evaluated using the  -Measure, where  = 1 weights the harmonic mean between precision and recall
equally, and the results are macro-averaged over all documents.</p>
        <p>All approaches are submitted on the TIRA platform [29], which allows participants to evaluate and
optimize their methods based on training, validation, and unseen test data. For the test data, blind
evaluation ensures that participants cannot optimize their approaches based on the test data.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Survey of Submissions</title>
      <p>We received 16 software submissions and 15 working note papers for the task of multi-author writing
style analysis in 2024. Below is a brief description of the submitted solutions.</p>
      <p>Lv et al. [30] leverage the decoder of LLaMA-3 to obtain vector representations of paragraph pairs,
subsequently using these representations to perform binary classification via a feed-forward network.
To increase the eficiency of their model training, they use a technique called low-rank adaptation.</p>
      <p>Lin et al. [31] use an ensemble of multiple transformer-based models (ReBERTa, DeBERTa, and
ERNIE) to solve the task. Crucially, to improve performance for the easy and medium datasets, where
topical variety within the documents is higher, they also perform a post-processing step based on
the semantic similarity of two consecutive paragraphs for those two datasets; paragraphs with a high
degree of semantic similarity are then deemed to have been written by the same author, irrespective of
the predictions obtained from the transformer ensemble.</p>
      <p>The submission of Ye et al. [32] utilizes continual learning to approach the task. Their goal is to
achieve a knowledge transfer across diferent dificulty levels, using learned progress prompts to do so.</p>
      <p>Huang and Kong [33] employ DeBERTa-v3 to fine-tune a model for this year’s task. To improve
the performance of the model, they use regularized dropout during the fine-tuning process. They also
perform early stopping during the training process to prevent the model from overfitting.</p>
      <p>The approach by Huang and Kong [34] employs models of the BERT family to solve the task. Like
most other participants, they fine-tuned the models on the training sets and then tested the performance
of various BERT-derived models on the validation set to decide on which model to use for the final
submission. Ultimately, they settled on DeBERTa for the easy and hard datasets, and on RoBERTa for
the medium dataset.</p>
      <p>Wu, Kong, and Ye [35] use RoBERTa to encode the positive and negative sample paragraph pairs.
They add a contrastive learning component to optimize the training process of RoBERTa that essentially
aims to reduce the cosine distance of positive paragraph pairs while increasing the distance of negative
paragraph pairs.</p>
      <p>Liu et al. [36] also employ contrastive learning for the encoding phase, using RoBERTa as the encoder.
For each pair of paragraphs, they form a feature matrix, consisting of the latent representations of the
two paragraphs, and the absolute distance between the two embeddings. The feature matrix is then fed
into a fully connected layer to compute the final prediction.</p>
      <p>Księżniak et al. [37] utilize RoBERTa and DeBERTa models for their solution. To give the models
additional information they could use to determine style changes, they augmented the texts of the
documents with tags containing stylometric features.</p>
      <p>Chen, Hand, and Yi [38] use RoBERTa and for the fine-tuning phase, they employ R-Drop
regularization to mitigate overfitting and to ensure consistency that the model, given identical inputs, computes
consistent predictions.</p>
      <p>Wu et al. [39] compared the performance of BERT, RoBERTa, and DistilBERT for task 1 and showed
that RoBERTa achieved the best results. Consequently, they used RoBERTa for the encoding and feed the
resulting pooled contextual features into a Virtual Softmax layer to perform a three-class classification
task, where the intuition behind introducing a third class is to enforce stricter boundary constraints
between the two original classes.</p>
      <p>Khan et al. [40] in a first step, performed weighted sampling on the data provided to achieve balanced
classes. They then compared RoBERTa, Electra, DeBerta, and Squeeze-Bert models, where RoBERTa was
performing best. To further enhance the performance, they augmented the provided data by swapping
all pairs of paragraphs between which no style change was detected and adding these paragraphs to
the training data.</p>
      <p>The approach by Sheykhlanet al. [41] makes use of fine-tuned transformer models, namely BERT,
RoBERTa, and ELECTRA, to detect style changes. They opted to use diferent combinations of models
depending on the dificulty of the dataset. For the easy dataset, they only used RoBERTa, while for the
medium and hard datasets, they used an ensemble of all three models.</p>
      <p>Sanjesh and Mangai [42] base their approach on latent representations of paragraphs by computing
embeddings on a set of stylometric features such as TF/IDF for character n-grams, stop word frequency,
character and word counts. These embeddings are then fed into a convolutional neural network and
Bi-directional LSTM layers, which are then combined in a dense layer.</p>
      <p>Liang and Lei [43] use GPT-3.5 as a teacher model that creates a dataset based on the provided
datasets by providing pairs of sentences to the model and then asking questions about the similarity of
topic, style, and vocabulary, and whether the sentences were written by the same author. The student
model employed is T5-small is then fine-tuned for the multi-author writing style analysis task.</p>
      <p>Liu, Chen, and Lv [44] leverage the Entropy-based Stability-Plasticity (ESP) method to tackle this
year’s task. ESP aims to balance stability and plasticity by restricting changes to the learning rate in
each layer based on entropy. As an encoder, the team used BERT.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation Results</title>
      <p>The results for all of this year’s submissions are shown in Table 1. The best result for each dificulty is
highlighted in bold; note thar the best result for each dificulty was achieved with diferent approaches.
For the easy dataset, both Ye et al. [32] and Huang and Kong [34] achieved first place with an F 1 of
0.991. For the medium dataset, the best result was obtained by Lv et al. [30] with an F1 of 0.887, while
the best result for the hard dataset was obtained by Lin et al. [31] with an F1 of 0.863.</p>
      <p>While there is still a clear diference in model performance between the three dificulty levels, the
results have converged significantly again this year, with higher scores for the medium and hard datasets
compared to last year, while the models on the easy dataset are already achieving near perfect scores.</p>
      <p>We also checked how the number of authors in a document afects the performance of the submitted
models for the medium and hard datasets. The results of this can be seen in Figure 1. We confirm the
same observation as in the previous two years: The performance of many submitted models on the hard
dataset, including the strongest submitted model, is better for documents written by three authors than
for those written by two authors. Most models then decrease in their performance again on documents
written by four authors, while the winning model maintains its performance for these documents.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In the 2024 edition of the multi-author writing style analysis task at PAN, the task was to identify
locations of writing style changes at the paragraph level. We provided participants with three datasets of
increasing thematic homogeneity and therefore dificulty. This year, we received 16 software submissions
and 15 working papers. The results obtained again show considerable progress compared to the results
of previous years.
[3] E. Stamatatos, Intrinsic Plagiarism Detection Using Character n-gram Profiles, in: B. Stein, P. Rosso,
E. Stamatatos, M. Koppel, E. Agirre (Eds.), SEPLN 2009 Workshop on Uncovering Plagiarism,
Authorship, and Social Software Misuse (PAN 09), Universidad Politécnica de Valencia and
CEURWS.org, 2009, pp. 38–46. URL: http://ceur-ws.org/Vol-502.
[4] M. Koppel, J. Schler, S. Argamon, Computational methods in authorship attribution, Journal of
the American Society for Information Science and Technology 60 (2009) 9–26.
[5] I. Bensalem, P. Rosso, S. Chikhi, Intrinsic plagiarism detection using n-gram classes, in:
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing
(EMNLP), Association for Computational Linguistics, Doha, Qatar, 2014, pp. 1459–1464. URL:
https://www.aclweb.org/anthology/D14-1153. doi:10.3115/v1/D14-1153.
[6] D. I. Holmes, The Evolution of Stylometry in Humanities Scholarship, Literary and Linguistic</p>
      <p>Computing 13 (1998) 111–117.
[7] R. Zheng, J. Li, H. Chen, Z. Huang, A Framework for Authorship Identification of Online
Messages: Writing-Style Features and Classification Techniques, Journal of the American Society for
Information Science and Technology 57 (2006) 378–393.
[8] M. Tschuggnall, G. Specht, Countering Plagiarism by Exposing Irregularities in Authors’ Grammar,
in: Proceedings of the European Intelligence and Security Informatics Conference (EISIC), IEEE,
Uppsala, Sweden, 2013, pp. 15–22.
[9] M. Tschuggnall, G. Specht, Automatic decomposition of multi-author documents using grammar
analysis, in: F. Klan, G. Specht, H. Gamper (Eds.), Proceedings of the 26th GI-Workshop Grundlagen
von Datenbanken, volume 1313 of CEUR Workshop Proceedings, CEUR-WS.org, 2014, pp. 17–22.</p>
      <p>URL: http://ceur-ws.org/Vol-1313.
[10] A. Glover, G. Hirst, Detecting Stylistic Inconsistencies in Collaborative Writing, Springer London,</p>
      <p>London, 1996, pp. 147–168. doi:10.1007/978-1-4471-1482-6_12.
[11] M. Koppel, N. Akiva, I. Dershowitz, N. Dershowitz, Unsupervised decomposition of a document
into authorial components, in: Proceedings of the 49th Annual Meeting of the Association
for Computational Linguistics: Human Language Technologies, Association for Computational
Linguistics, Portland, Oregon, USA, 2011, pp. 1356–1364. URL: https://www.aclweb.org/anthology/
P11-1136.
[12] M. Koppel, N. Akiva, I. Dershowitz, N. Dershowitz, Unsupervised decomposition of a document
into authorial components, in: D. Lin, Y. Matsumoto, R. Mihalcea (Eds.), The 49th Annual Meeting
of the Association for Computational Linguistics: Human Language Technologies, Proceedings of
the Conference, 19-24 June, 2011, Portland, Oregon, USA, The Association for Computer Linguistics,
2011, pp. 1356–1364. URL: http://www.aclweb.org/anthology/P11-1136.
[13] N. Akiva, M. Koppel, Identifying Distinct Components of a Multi-author Document, in: N. Memon,
D. Zeng (Eds.), 2012 European Intelligence and Security Informatics Conference, EISIC 2012, IEEE
Computer Society, 2012, pp. 205–209. URL: https://doi.org/10.1109/EISIC.2012.16. doi:10.1109/
EISIC.2012.16.
[14] N. Akiva, M. Koppel, A Generic Unsupervised Method for Decomposing Multi-Author Documents,</p>
      <p>JASIST 64 (2013) 2256–2264. URL: https://doi.org/10.1002/asi.22924. doi:10.1002/asi.22924.
[15] A. Rexha, S. Klampfl, M. Kröll, R. Kern, Towards a more fine grained analysis of scientific authorship:
Predicting the number of authors using stylometric features, in: P. Mayr, I. Frommholz, G. Cabanac
(Eds.), Proceedings of the Third Workshop on Bibliometric-enhanced Information Retrieval
colocated with the 38th European Conference on Information Retrieval (ECIR 2016), volume 1567 of
CEUR Workshop Proceedings, CEUR-WS.org, 2016, pp. 26–31. URL: http://ceur-ws.org/Vol-1567.
[16] E. Stamatatos, M. Tschuggnall, B. Verhoeven, W. Daelemans, G. Specht, B. Stein, M. Potthast,
Clustering by Authorship Within and Across Documents, in: Working Notes Papers of the
CLEF 2016 Evaluation Labs, CEUR Workshop Proceedings, CLEF and CEUR-WS.org, 2016. URL:
http://ceur-ws.org/Vol-1609/.
[17] M. Tschuggnall, E. Stamatatos, B. Verhoeven, W. Daelemans, G. Specht, B. Stein, M. Potthast,
Overview of the Author Identification Task at PAN 2017: Style Breach Detection and Author
Clustering, in: L. Cappellato, N. Ferro, L. Goeuriot, T. Mandl (Eds.), Working Notes Papers of the
CLEF 2017 Evaluation Labs, volume 1866 of CEUR Workshop Proceedings, CEUR-WS.org, 2017.</p>
      <p>URL: http://ceur-ws.org/Vol-1866/.
[18] M. Kestemont, M. Tschuggnall, E. Stamatatos, W. Daelemans, G. Specht, B. Stein, M. Potthast,
Overview of the Author Identification Task at PAN-2018: Cross-domain Authorship Attribution
and Style Change Detection, in: L. Cappellato, N. Ferro, J.-Y. Nie, L. Soulier (Eds.), Working
Notes Papers of the CLEF 2018 Evaluation Labs, volume 2125 of CEUR Workshop Proceedings,
CEUR-WS.org, 2018. URL: https://ceur-ws.org/Vol-2125/invited_paper_2.pdf.
[19] E. Zangerle, M. Tschuggnall, G. Specht, M. Potthast, B. Stein, Overview of the Style Change
Detection Task at PAN 2019, in: L. Cappellato, N. Ferro, D. Losada, H. Müller (Eds.), CLEF 2019
Labs and Workshops, Notebook Papers, CEUR-WS.org, 2019. URL: http://ceur-ws.org/Vol-2380/
paper_243.pdf.
[20] E. Zangerle, M. Mayerl, G. Specht, M. Potthast, B. Stein, Overview of the Style Change Detection
Task at PAN 2020, in: L. Cappellato, C. Eickhof, N. Ferro, A. Névéol (Eds.), CLEF 2020 Labs and
Workshops, Notebook Papers, CEUR-WS.org, 2020. URL: https://ceur-ws.org/Vol-2696/paper_256.
pdf.
[21] E. Zangerle, M. Mayerl, , M. Potthast, B. Stein, Overview of the Style Change Detection Task
at PAN 2021, in: G. Faggioli, N. Ferro, A. Joly, M. Maistro, F. Piroi (Eds.), CLEF 2021 Labs and
Workshops, Notebook Papers, CEUR-WS.org, 2021, pp. 1760–1771. URL: https://ceur-ws.org/
Vol-2936/paper-148.pdf.
[22] E. Zangerle, M. Mayerl, M. Potthast, B. Stein, Overview of the Style Change Detection Task at PAN
2022, in: G. Faggioli, N. Ferro, A. Hanbury, M. Potthast (Eds.), CLEF 2022 Labs and Workshops,
Notebook Papers, CEUR-WS.org, 2022. URL: http://ceur-ws.org/Vol-3180/paper-186.pdf.
[23] E. Zangerle, M. Mayerl, M. Potthast, B. Stein, Overview of the Multi-Author Writing Style Analysis
Task at PAN 2023, in: M. Aliannejadi, G. Faggioli, N. Ferro, M. Vlachos (Eds.), Working Notes of
the Conference and Labs of the Evaluation Forum (CLEF 2023), volume 3497 of CEUR Workshop
Proceedings, 2023, pp. 2513–2522. URL: https://ceur-ws.org/Vol-3497/paper-201.pdf.
[24] D. Zlatkova, D. Kopev, K. Mitov, A. Atanasov, M. Hardalov, I. Koychev, P. Nakov, An Ensemble-Rich
Multi-Aspect Approach for Robust Style Change Detection, in: L. Cappellato, N. Ferro, J.-Y. Nie,
L. Soulier (Eds.), CLEF 2018 Evaluation Labs and Workshop – Working Notes Papers, CEUR-WS.org,
2018. URL: https://ceur-ws.org/Vol-2125/paper_142.pdf.
[25] A. Iyer, S. Vosoughi, Style Change Detection Using BERT—Notebook for PAN at CLEF 2020, in:
L. Cappellato, C. Eickhof, N. Ferro, A. Névéol (Eds.), CLEF 2020 Labs and Workshops, Notebook
Papers, CEUR-WS.org, 2020. URL: http://ceur-ws.org/Vol-2696/.
[26] Z. Zhang, Z. Han, L. Kong, X. Miao, Z. Peng, J. Zeng, H. Cao, J. Zhang, Z. Xiao, X. Peng, Style Change
Detection Based On Writing Style Similarity—Notebook for PAN at CLEF 2021, in: G. Faggioli,
N. Ferro, A. Joly, M. Maistro, F. Piroi (Eds.), CLEF 2021 Labs and Workshops, Notebook Papers,
CEUR-WS.org, 2021. URL: http://ceur-ws.org/Vol-2936/paper-198.pdf.
[27] T.-M. Lin, C.-Y. Chen, Y.-W. Tzeng, L.-H. Lee, Ensemble Pre-trained Transformer Models for Writing
Style Change Detection, in: G. Faggioli, N. Ferro, A. Hanbury, M. Potthast (Eds.), CLEF 2022
Labs and Workshops, Notebook Papers, CEUR-WS.org, 2022. URL: http://ceur-ws.org/Vol-3180/
paper-210.pdf.
[28] H. Chen, Z. Han, Z. Li, Y. Han, A Writing Style Embedding Based on Contrastive Learning for
Multi-Author Writing Style Analysis, in: M. Aliannejadi, G. Faggioli, N. Ferro, M. Vlachos (Eds.),
Working Notes of CLEF 2023 - Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2023,
pp. 2562–2567. URL: https://ceur-ws.org/Vol-3497/paper-206.pdf.
[29] M. Potthast, T. Gollub, M. Wiegmann, B. Stein, TIRA Integrated Research Architecture, in: N. Ferro,
C. Peters (Eds.), Information Retrieval Evaluation in a Changing World, The Information Retrieval
Series, Springer, Berlin Heidelberg New York, 2019. doi:10.1007/978-3-030-22948-1_5.
[30] J. Lv, Y. Yi, H. Qi, Team Fosu-stu at PAN: Supervised fine-tuning of large language models for Multi
Author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.),
Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[31] T. Lin, Y. Wu, L. Lee, Team NYCU-NLP at PAN 2024: Integrating Transformers with Similarity
Adjustments for Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková,
A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation
Forum, CEUR-WS.org, 2024.
[32] Z. Ye, Y. Zhong, C. Huang, L. Kong, Team no-999 at PAN: Continual Transfer Learning with
Progress Prompt for Multi-Author Writing Style Analysis", in: G. Faggioli, N. Ferro, P. Galuščáková,
A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation
Forum, CEUR-WS.org, 2024.
[33] Z. Huang, L. Kong, Team huangzhijian at PAN: DeBERTa-v3 with R-Drop Regularization for
Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera
(Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org,
2024.
[34] Y. Huang, L. Kong, Team text-understanding-and-analysi at PAN: Utilizing BERT Series
Pretraining Model for Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková,
A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation
Forum, CEUR-WS.org, 2024.
[35] Q. Wu, L. Kong, Z. Ye, Team bingezzzleep at PAN: A Writing Style Change Analysis Model Based
on RoBERTa Encoding and Contrastive Learning for Multi-Author Writing Style Analysis, in:
G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024
Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[36] C. Liu, Z. Han, H. Chen, Q. Hu, Team liuc0757 at PAN: A Writing Style Embedding Method
Based on Contrastive Learning for Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro,
P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of
the Evaluation Forum, CEUR-WS.org, 2024.
[37] E. Ksi¸eżniak, K. W¸ecel, M. Sawiński, Team OpenFact at PAN 2024: Fine-Tuning BERT Models with
Stylometric Enhancements, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.),
Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[38] Z. Chen, Y. Han, Y. Yi, Team chen at PAN: Integrating R-Drop and Pre-trained Language Model for
Multi-author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera
(Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org,
2024.
[39] B. Wu, Y. Han, K. Yan, H. Qi, Team baker at PAN: Enhancing Writing Style Change Detection
with Virtual Softmax, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.), Working
Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[40] A. Khan, M. Rai, K. Khan, S. Shah, F. Alvi, A. Samad, Team Gladiators at PAN: Improving
Author Identification: A Comparative Analysis of Pre-Trained Transformers for Multi-Author
Classification, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes
of CLEF 2024 - Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[41] M. Sheykhlan, S. Abdoljabbar, M. Mahmoudabad, Team karami-sh at PAN: Transformer-based
Ensemble Learning for Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro, P. Galuščáková,
A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation
Forum, CEUR-WS.org, 2024.
[42] R. Sanjesh, A. Mangai, Team riyasanjesh at PAN: Multi-feature with CNN and Bi-LSTM Neural
Network approach to Style Change Detection, in: G. Faggioli, N. Ferro, P. Galuščáková, A. G. S.
de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum,
CEUR-WS.org, 2024.
[43] X. Liang, H. Lei, Team lxflcl66666 at PAN: Fine-Tuned Reasoning for Writing Style Analysis, in:
G. Faggioli, N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024
Conference and Labs of the Evaluation Forum, CEUR-WS.org, 2024.
[44] X. Liu, H. Chen, J. Lv, Team foshan-university-of-guangdong at PAN: Adaptive
EntropyBased Stability-Plasticity for Multi-Author Writing Style Analysis, in: G. Faggioli, N. Ferro,
P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 - Conference and Labs of
the Evaluation Forum, CEUR-WS.org, 2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Karaś</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Śpiewak</surname>
          </string-name>
          , P. Sobecki, OPI-JSA at CLEF 2017:
          <article-title>Author Clustering and Style Breach Detection-Notebook for PAN at CLEF 2017</article-title>
          , in: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
          </string-name>
          , T. Mandl (Eds.),
          <article-title>CLEF 2017 Evaluation Labs</article-title>
          and Workshop - Working Notes Papers, CEUR-WS.org,
          <year>2017</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-1866/.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Khan</surname>
          </string-name>
          , Style Breach Detection:
          <article-title>An Unsupervised Detection Model-Notebook for PAN at CLEF 2017</article-title>
          , in: L.
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
          </string-name>
          , T. Mandl (Eds.),
          <article-title>CLEF 2017 Evaluation Labs</article-title>
          and Workshop - Working Notes Papers, CEUR-WS.org,
          <year>2017</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-1866/.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>