<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Analysis in English Texts Based on Graph Convolutional Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Abeer Saad Alsheddi</string-name>
          <email>asalsheddi@imamu.edu.sa</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohamed El Bachir Menai</string-name>
          <email>menai@ksu.edu.sa</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer Science Department, Imam Mohammad Ibn Saud Islamic University</institution>
          ,
          <addr-line>Riyadh</addr-line>
          ,
          <country country="SA">Saudi Arabia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Computer Science Department, King Saud University</institution>
          ,
          <addr-line>Riyadh</addr-line>
          ,
          <country country="SA">Saudi Arabia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The Multi-Author Writing Style Analysis (MAWSA) task asks to find the locations of writing style changes at diferent text levels. This task can assist in other applications such as plagiarism, security, and commerce. 2020, existing MAWSA models have commonly represented any boundary between two consecutive segments by joining them. The representations of these joined segments then serve as the input for these models. This join may lose style features within each segment. In this paper, the proposed method exploits relationships between segments using Graph Convolutional Networks (GCNs). Boundaries and segment representations are depicted independently. The PAN 2025 dataset is provided at three diferent levels of topic distributions: easy, medium, and hard, while changes appear on the sentence level. The trained model, named STAR-GCN-MAWSA, achieved an  1-score of 0.857, 0.764, and 0.662 for easy, medium, and hard MAWSA instances on validation sets, respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>Convolutional</kwd>
        <kwd>Style change detection</kwd>
        <kwd>Multi-author Analysis</kwd>
        <kwd>Graph convolutional networks</kwd>
        <kwd>Pretrained models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>PAN1 organizes a series of scientific competitions to promote research on stylometry and digital text
forensics. It has provided a Multi-Author Writing Style Analysis (MAWSA) task since 2017. This
task focuses on diferentiating author styles within multi-authored text documents without providing
comparison documents. It segments a text document. It then examines the boundary between each
segment to determine whether it separates two segments written by the same author. For example, if a
document is segmented into five segments  1 to  5, there are four boundaries located between these
ifve segments  1 to  4. Developing MAWSA models can assist other practical applications such as
plagiarism, security, and commerce. In plagiarism, MAWSA solutions can suggest potential plagiarism
cases by identifying changes in writing style without comparing the suspected and source documents.
In security measures, unauthorized modifications to sensitive documents can be identified to fortify the
security systems. In commerce, the coherence of writing style can be improved for proofreaders and
institutions by minimizing variations in writing style to adhere to a single style in their documents.</p>
      <p>
        The previous PAN editions aimed to tackle the MAWSA task from diferent aspects by proposing
diferent levels of subtasks, which can be categorized into four subtasks. The first subtask determined
single/multi-author document, which was provided in MAWSA 2017 - MAWSA 2022 [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5 ref6">1, 2, 3, 4, 5, 6</xref>
        ].
The second subtask detected change positions (boundaries) on the sentence level in MAWSA 2017 and
MAWSA 2022 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or on the paragraph level in MAWSA 2020 - MAWSA 2024 [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ]. The third subtask
identified the actual number of authors who wrote a given document in MAWSA 2019 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The last
subtask considered the attribution that assigned all segments uniquely to their respective authors in
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
      <p>
        MAWSA 2021 and MAWSA 2022 [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. In this year, MAWSA 2025 [
        <xref ref-type="bibr" rid="ref7">7, 8</xref>
        ] is related to the second subtask,
detecting change positions on the sentence level, while focusing on topic diversity in datasets.
      </p>
      <p>Since 2020, most existing models represent boundaries between segments by concatenating these
segments. For example, the boundary   concatenates two segments   and  +1 to form one pair    +1 .
This input pair is then represented by using the representations of the pair    +1 as input in most
existing models. This concatenation can eliminate the need to explicitly work on boundary features.
However, it loses segment representations. In other words, this concatenation does not preserve the
representation of each segment alone through the processing within models. Thus, segments cannot
be retrieved at the end of the processing. For example, author attribution and author counting were
studied in MAWSA 2019, MAWSA 2021, and MAWSA 2022, which are based on segment representations
themselves. This motivates us to close this gap.</p>
      <p>In addition, comprehending relationships between textual segments, such as words and sentences,
would enhance the detection of writing styles. Graph-based solutions take a graph as input, trying
to involve structural properties within the data. Graph Neural Networks (GNNs) extend existing
neural networks to operate directly on graph-structured data [9]. Recently, GNN models have achieved
promising results for some Natural Language Processing (NLP) tasks, such as an authorship verification
task that determines whether an unknown text was written by a specific author [ 10, 11, 12] and semantic
relationship tasks that analyze semantic relations between textual segments [13, 14, 15, 16, 17, 18, 19,
20, 21, 22, 23, 24, 25]. To the best of our knowledge, no existing GNN-based solution is available for the
MAWSA task.</p>
      <p>In this study, we participated in the MAWSA 2025 task by representing boundaries as standalone
matrices while preserving the segment representations surrounding these boundaries. The boundary
representations are learnable and updated within layers. This work also explores Graph Convolutional
Networks (GCNs) [26] for MAWSA. It is considered the first to leverage the characteristics of graph
neural networks to detect style changes. In this work, graphs indicate documents, nodes represent
sentences, and edges state boundaries. Moreover, sentence features were extracted using STAR [27]. It
is a recent pre-trained model trained on authorship representations, which is more related to MAWSA.</p>
      <p>The remaining parts of this paper are organized as follows. Section 2 describes the task and the
provided dataset in MAWSA 2025. Section 3 investigates related work for the task. Sections 4 and 5
describe the proposed approach and present its results. Conclusions are raised in Section 6.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Description and Dataset</title>
      <p>The MAWSA 2025 task asks to determine whether the writing style changes on the sentence level in a
given document. This edition pays more attention to topic diversity in datasets. Therefore, it provides
the datasets with three diferent levels of topic diversity to decrease the use of topic information in
identifying style changes [8]:</p>
      <sec id="sec-2-1">
        <title>Dataset 1 (Easy): The sentences in a document cover various topics.</title>
        <p>Dataset 2 (Medium): The sentences in a document cover fewer topics than the easy level.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Dataset 3 (Hard) All the sentences in a document cover the same topic.</title>
        <p>The results of simple statistical analysis are shown in Table 1. Each dataset was split into training,
validation, and test sets. The training and validation sets are available with ground truth labels to
train and optimize proposed models. The test set is hidden until the end of the competition, and it is
not publicly available so far. The average length of documents is measured as the average number of
sentences per document. The average length of sentences is measured as the average number of words
per sentence. The average number of writing style changes and the percentage of these changes are
measured per document.</p>
        <p>The input files in the dataset contain a row of English text *.txt. The expected output is a list of binary
values representing the change of writing style within a document. Value ’0’ indicates that the same</p>
        <sec id="sec-2-2-1">
          <title>Dataset</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>Dataset 1</title>
          <p>(Easy)
Dataset 2
(Medium)
Dataset 3
(Hard)</p>
          <p>Split
train
validation</p>
          <p>train
validation</p>
          <p>train
validation
author writes the two consecutive sentences and then has the same style. While Value ’1’ suggests that
their authors are diferent and have unique writing styles.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Related Works</title>
      <p>This section provides an overview of the methods incorporated in previous MAWSA works. According
to the conducted review, three methods were adopted: statistical, classical machine learning, and deep
neural networks [28].</p>
      <p>In statistical methods, models were developed by selecting stylistic features, followed by applying
statistical methods without training their models [29, 30, 31]. Khan [29, 30] defined a measure for each
type of handcrafted feature, assuming a style has changed if the score is less than a threshold. Karas
et al. [31] adopted a distribution test called the Wilcoxon Signed Rank Test [32] to predict the style
changes.</p>
      <p>In machine learning methods, models are obtained based on either supervised or unsupervised
learning. Most of supervised-based works relied on the logistic regression and the random forest
algorithms [33, 34, 35], whereas Support Vector Machine outperformed them in other works proposed
MAWSA 2018 [36]. The unsupervised learning-based works mostly utilized clustering documents based
on the similarity of their writing styles using the K-means clustering algorithm [34, 37, 38, 39, 40] with
the cosine similarity function that outperformed Jaccard and Dice functions [37].</p>
      <p>In deep neural network methods, features represent whole documents instead of selecting specific
features. Documents in these works were fed into a CNN model [41], a Siamese neural network of one
or two BiLSTM [36, 42] layers. Other works used pretrained models, such as ELECTRA [43] and BERT
with a CNN layer [44, 45], an MLM head [46], or feedforward neural networks [47, 48].</p>
    </sec>
    <sec id="sec-4">
      <title>4. Proposed Method</title>
      <p>The proposed method is illustrated by Figure 1. It is based on a deep neural network architecture
comprising four layers that extract more non-local features. The following paragraphs provide a more
detailed description of them.</p>
      <p>Features: The input documents were segmented into sentences. This research uses the pretrained
model Style Transformer for Authorship Representations (STAR) [27] to represent sentences. STAR
characterizes writing style in social media and trained for learning authorship representations. A
pretrained model is a transformer with trained parameters and saved values from a large dataset. These
models can adapt their parameters to better suit a particular task by retraining some or all. Pretrained
models have provided high results in SOTA models.   ×(0) (0) in in Figure 1 indicates to the initial
node representations. All the trainable parameters of STAR were frozen. Thus, no fine-tuning was
performed on their parameters for extracting embeddings. This freezing allowed us to assess the models’
capabilities rigorously within the constraints of our experimental setup.</p>
      <p>Graphs: After that, the representations are structured as graphs. Each document is treated as a
disjoint subgraph because the task does not look for a common writing style between two documents.
Thus, there is no direct relationship between any two documents, so they can be treated as a disjoint
subgraph. Each node represents a sentence. Edges connect the preceding to the succeeding sentences.
Thus, edges concern the boundaries between sentences. The edge representations are the writing styles
of consecutive sentences. The input graph  (0) has an empty set of edge representations 
These representations are updated across layers and then classified for the MAWSA task. As shown in
(0)
 −1×
(0).
only consecutive paragraphs.</p>
      <p>GCNs was introduced by Kipf and Welling [26] as one of the early GNN models. GCNs are suitable for
learning the representation of nodes in the input graph. GCNs help capture the relationships between
nodes, in our case, between the sentence styles Equation 1 shows GCN’s message-passing schema. The
messages from all neighbors  ()
are normalized by the degrees of the neighbor  ∈  ()
and the target
 nodes. These messages are then summed to aggregate them. The aggregated messages are combined
with the current target node representations to update the latter. Although GCNs are widely used for
diverse NLP tasks [16, 17, 13, 24, 18, 21, 14, 25, 15], GCNs focus on node representations and can handle
edge weights within the adjacency matrices as shallow edge representations.</p>
      <p>x
() =</p>
      <p>∑
∈ ()∪{}</p>
      <p>1
√deg() ⋅ √deg()
⋅ (W⊤ ⋅ x(−1) )</p>
      <p>Edge representation: Every layer  outputs new edge representations. These representations
aggregate three values. First, the edge representations from the previous layer  (−1) , whereas the first
layer receives an empty set. Second, the node representations  () are generated from GCNs() , where
GCNs() receive the previous node representations  (−1) . Third, the initial node representations,  (0),
are added to alleviate the over-smoothing issue. The symbol ⊕ in Figure 1 indicates to sum these
three values to generate edge representations in the current layer,  () . Equation (2) shows how the
edge representations e</p>
      <p>() can be updated, where  (⋅) represents a nonlinear activation function, 
represents the summation operation as the aggregation function,   and   indicate the representations
of end nodes  and  of edge   ( ∈  ()</p>
      <p>), and   is a learnable parameter in the layer  that adjusts
the dimension of output representation vectors. The edge representations in each layer are updated
according to the new node representation in the same layer  () .</p>
      <p>e
() = Sum(W e

(−1) ,  (

 (x−1 , x−1 )),W0x0).</p>
      <p>Classification:</p>
      <p>The output graph  (4) contains edge representations 
tations are classified by a single fully connected (FC) layer. It is followed by the activation function
Sigmoid, and then are classified using a threshold of 0.5 to round outputs to 0 and 1.
(4)
 −1×
(4). These
represen(1)
(2)</p>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation</title>
      <p>5.1. Experiment Settings
Setup: Each input document is used as one batch. Dropout rates of 0.5, warmup rates of 0.1, learning
rate of 2e-5, and 20 epochs were used during the training. The experiments were conducted using a
personal computer with the following specifications:
• CPU: Intel(R) i7 processor up to 5.60 GHz, 64-bit
• GPU: ASUS TUF RTX 4090 24GB OC GAMING.
• RAM: 64 GB (2x32 GB) DDR5 5600 Mhz
• Programming language: Python with the seed of 42 and the PyTorch framework.</p>
      <p>Encoder: Three candidates were selected. They are three pretrained models used with their default
configuration. BERT 2 had the top usage from 2020 to 2022 [44, 46, 45, 47, 48, 49, 50], RoBERTa 3 had
the top usage in the latest two editions 2023 and 2024 [51, 52, 53, 54, 55, 56, 56, 57, 51, 58], and STAR4 is
used as the third pretrained model. This study used 256 tokens as the maximum length. Any sentence
exceeding this length will be truncated. The special token [CLS] was used to represent the entire
sentence. All the models were developed under the same setting.</p>
      <p>Evaluation metric: The performance of each model is measured using the F1-score5 metric. The
macro-averaged computes each class’s average separately, change or not change, and returns the average
without considering the proportion of each class in the dataset.
5.2. Result and Discussion
2https://huggingface.co/google-bert/bert-base-uncased
3https://huggingface.co/FacebookAI/roberta-base
4https://huggingface.co/AIDA-UPM/star
5https://scikit-learn.org/1.5/modules/generated/sklearn.metrics.f1_score.html
less than a quarter. This small percentage leads models to train on almost the not change case, making
it harder to detect change in the validation set. In particular, the number of changes has diminished as
the datasets become more complicated. This bias also appears in some previous editions of MAWSA
datasets [28].</p>
      <p>It is important to bear in mind the possible bias, specifically in Datasets 1 and 2. This case may
be related to the topic distributions in the training and the validation sets. The diference in the
distributions may guide models to train on specific styles more than others, especially with a small size
of the validation set. Further research is needed to understand the relationship between the distribution
of topics and writing styles.</p>
      <p>Ablation experiments were conducted to evaluate the components of the edge features, which were
added cumulatively in the experiments. Table 3 shows the results obtained from the three datasets. First,
the basic four GCN layers were developed as the baseline models, and their edge representations were
measured by summing the representations of the two end nodes extracted from the fourth GCN layer.
Second, a warmup mechanism optimized the model performance. Third, adding edge representations
obtained from previous layers helped adjust them across layers. Fourth, initial node information
was aggregated into edge representations. Table 3 shows that the fusion of both initial node and
edge representations enhances the learning of edge representations for MAWSA and can mitigate the
over-smoothing issue.</p>
      <p>Table 4 shows the results on the test sets shown on the Tira platform. The results obtained from the
validation and test sets have revealed some intriguing disparities, despite both sets being withheld during
the training process. While the validation results suggest that STAR-GCN-MAWSA performs acceptably,
especially on easy instances, the test results have not mirrored this stability. This inconsistency between
the results may be due to various factors. One plausible explanation could be the presence of data
distribution diferences between the validation and test sets, which leads to an increase in the model’s
sensitivity during evaluation. Investigating these discrepancies further in the future is essential through
an analysis of the data distribution to ensure consistency between the sets and help achieve more
stability.</p>
      <p>Beyond the conclusion of the oficial competition, our eforts to enhance the GNN-based solution
for MAWSA continued. The latest advancement in this ongoing work has been achieved through
the integration of an alternative GNN module: EdgeConv [60]. This specific architecture was chosen
for its ability to incorporate edge representations directly into the node message-passing mechanism,
thereby enabling a richer understanding of local graph structures and relationships. This
EdgeConvbased solution yielded improved performance on the validation set, as detailed in Table 5. A more
comprehensive description of this advanced EdgeConv-based approach and its implementation can be
found in our recent work [61].</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>This paper set out to address the MAWSA 2025 task, which focuses on the topic diversity in datasets on
the sentence level. The proposed solution mainly considers to integrate GCNs to address boundary style
features independently. This method updates edge features across layers while preserving sentence
representations. One of the findings from this study is that the STAR-GCN-MAWSA model outperformed
the two models, BERT-GCN-MAWSA and RoBERTa-GCN-MAWSA. The other major finding is that
aggregating three components for representing boundary styles achieved high results for the MAWSA
task. As future work, we plan to investigate the static bias that may occur during model training and
its impact on the model’s performance.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <sec id="sec-7-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
        <p>Detoxification, Multi-Author Writing Style Analysis, and Generative Plagiarism Detection, in:
J. C. de Albornoz, J. Gonzalo, L. Plaza, A. G. S. de Herrera, J. Mothe, F. Piroi, P. Rosso, D. Spina,
G. Faggioli, N. Ferro (Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction.
Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF 2025),
Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York, 2025.
[8] E. Zangerle, M. Mayerl, M. Potthast, B. Stein, Overview of the Multi-Author Writing Style Analysis
Task at PAN 2025, in: G. Faggioli, N. Ferro, P. Rosso, D. Spina (Eds.), Working Notes of CLEF 2025
– Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org,
2025.
[9] F. Scarselli, M. Gori, Ah Chung Tsoi, M. Hagenbuchner, G. Monfardini, The Graph Neural Network</p>
        <p>Model, IEEE Transactions on Neural Networks 20 (2009) 61–80.
[10] D. Embarcadero-Ruiz, H. Gómez-Adorno, A. Embarcadero-Ruiz, G. Sierra, Graph-Based Siamese
Network for Authorship Verification, in: Working Notes of CLEF 21- Conference and Labs of
the Evaluation Forum, volume 2936 of CEUR Workshop Proceedings, CEUR-WS.org, Bucharest,
Romania, 2021, p. 11.
[11] D. Embarcadero-Ruiz, H. Gómez-Adorno, A. Embarcadero-Ruiz, G. Sierra, Graph-Based Siamese
Network for Authorship Verification, in: Working Notes of CLEF 22- Conference and Labs of
the Evaluation Forum, volume 3180 of CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy,
2022, p. 277.
[12] A. Valdez-Valenzuela, J. A. Martinez-Galicia, H. Gomez-Adorno, Heterogeneous-Graph
Convolutional Network for Authorship Verification, in: Working Notes of CLEF 2023 - Conference and
Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org, Thessaloniki, Greece,
2023, p. 8.
[13] Y. Zhang, P. Qi, C. D. Manning, Graph Convolution over Pruned Dependency Trees Improves
Relation Extraction, in: Proceedings of the 2018 Conference on Empirical Methods in Natural
Language Processing, Association for Computational Linguistics, Brussels, Belgium, 2018, pp.
2205–2215.
[14] B. Yu, X. Mengge, Z. Zhang, T. Liu, W. Yubin, B. Wang, Learning to Prune Dependency Trees with
Rethinking for Neural Relation Extraction, in: Proceedings of the 28th International Conference
on Computational Linguistics, International Committee on Computational Linguistics, Barcelona,
Spain (Online), 2020, pp. 3842–3852.
[15] B. Li, Y. Fan, Y. Sataer, Z. Gao, Y. Gui, Improving Semantic Dependency Parsing with Higher-Order</p>
        <p>Information Encoded by Graph Neural Networks, Applied Sciences 12 (2022) 16.
[16] Y. Hu, H. Shen, W. Liu, F. Min, X. Qiao, K. Jin, A Graph Convolutional Network With Multiple</p>
        <p>Dependency Representations for Relation Extraction, IEEE Access 9 (2021) 81575–81587.
[17] K. Sun, R. Zhang, Y. Mao, S. Mensah, X. Liu, Relation Extraction with Convolutional Network over
Learnable Syntax-Transport Graph, Proceedings of the AAAI Conference on Artificial Intelligence
34 (2020) 8928–8935.
[18] L. Zhou, T. Wang, H. Qu, L. Huang, Y. Liu, A Weighted GCN with Logical Adjacency Matrix for</p>
        <p>Relation Extraction, Santiago de Compostela (2020) 8.
[19] A. Mandya, D. Bollegala, F. Coenen, Graph Convolution over Multiple Dependency Sub-graphs
for Relation Extraction, in: Proceedings of the 28th International Conference on Computational
Linguistics, International Committee on Computational Linguistics, Barcelona, Spain (Online),
2020, pp. 6424–6435.
[20] Z. Jin, Y. Yang, X. Qiu, Z. Zhang, Relation of the Relations: A New Paradigm of the Relation</p>
        <p>Extraction Problem, CoRR abs/2006.03719 (2020) 11.
[21] Y. Tian, G. Chen, Y. Song, X. Wan, Dependency-driven Relation Extraction with Attentive Graph
Convolutional Networks, in: Proceedings of the 59th Annual Meeting of the Association for
Computational Linguistics and the 11th International Joint Conference on Natural Language
Processing (Volume 1: Long Papers), Association for Computational Linguistics, Online, 2021, pp.
4458–4471.
[22] K. Zhao, H. Xu, Y. Cheng, X. Li, K. Gao, Representation Iterative Fusion Based on Heterogeneous
Graph Neural Network for Joint Entity and Relation Extraction, Knowledge-Based Systems 219
(2021) 9.
[23] P. Liu, L. Wang, Q. Zhao, H. Chen, Y. Feng, X. Lin, L. He, ECNU_ica_1 SemEval-2021 Task
4: Leveraging Knowledge-enhanced Graph Attention Networks for Reading Comprehension of
Abstract Meaning, in: Proceedings of the 15th International Workshop on Semantic Evaluation
(SemEval-2021), Association for Computational Linguistics, Online, 2021, pp. 183–188.
[24] Z. Guo, Y. Zhang, W. Lu, Attention Guided Graph Convolutional Networks for Relation Extraction,
in: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics,
Association for Computational Linguistics, Florence, Italy, 2019, pp. 241–251.
[25] Z. Li, Y. Sun, J. Zhu, S. Tang, C. Zhang, H. Ma, Improve Relation Extraction with Dual
Attention</p>
        <p>Guided Graph Convolutional Networks, Neural Computing and Applications 33 (2021) 1773–1784.
[26] T. N. Kipf, M. Welling, Semi-Supervised Classification with Graph Convolutional Networks,</p>
        <p>Technical Report arXiv:1609.02907, arXiv, 2017.
[27] J. Huertas-Tato, A. Martín, D. Camacho, Understanding writing style in social media with a
supervised contrastively pre-trained transformer, Knowledge-Based Systems 296 (2024) 12.
[28] A. S. Alsheddi, M. E. B. Menai, Writing Style Change Detection: State of the Art, Challenges, and</p>
        <p>Research Opportunities, Artificial Intelligence Review (in press) (2025) 60.
[29] J. A. Khan, Style Breach Detection: An Unsupervised Detection Model, in: CLEF 2017 Labs and
Workshops, Notebook Papers, volume 1866 of CEUR Workshop Proceedings, CEUR-WS.org, Dublin,
Ireland, 2017, p. 10.
[30] J. A. Khan, A Model for Style Change Detection at a Glance, in: CLEF 2018 Labs and Workshops,
Notebook Papers, volume 2125 of CEUR Workshop Proceedings, CEUR-WS.org, Avignon, France,
2018, p. 8.
[31] D. Karas, M. Spiewak, S. Piotr, OPI-JSA at CLEF 2017: Author Clustering and Style Breach
Detection, in: CLEF 2017 Labs and Workshops, Notebook Papers, volume 1866 of CEUR Workshop
Proceedings, CEUR-WS.org, Dublin, Ireland, 2017, p. 12.
[32] K. M. Ramachandran, C. P. Tsokos, Mathematical statistics with applications in R, third ed., Elsevier,</p>
        <p>Philadelphia, 2020.
[33] R. Singh, J. Weerasinghe, R. Greenstadt, Writing Style Change Detection on Multi-Author
Documents, in: CLEF 2021 Labs and Workshops, Notebook Papers, volume 2936 of CEUR Workshop
Proceedings, CEUR-WS.org, Bucharest, Romania, 2021, p. 9.
[34] F. Alvi, H. Algafri, N. Alqahtani, Style Change Detection using Discourse Markers, in: CLEF 2022
Labs and Workshops, Notebook Papers, volume 3180 of CEUR Workshop Proceedings, CEUR-WS.org,
Bologna, Italy, 2022, p. 6.
[35] K. Sa n, A. Ogaltsov, Detecting a Change of Style Using Text Statistics, in: CLEF 2018 Labs
and Workshops, Notebook Papers, volume 2125 of CEUR Workshop Proceedings, CEUR-WS.org,
Avignon, France, 2018, p. 6.
[36] S. Nath, Style Change Detection using Siamese Neural Networks, in: CLEF 2021 Labs and
Workshops, Notebook Papers, volume 2936 of CEUR Workshop Proceedings, CEUR-WS.org, Bucharest,
Romania, 2021, p. 11.
[37] M. Elamine, S. Mechti, L. H. Belguith, An Unsupervised Method for Detecting Style Breaches
in a Document, in: 2019 IEEE/ACS 16th International Conference on Computer Systems and
Applications (AICCSA), IEEE, Abu Dhabi, United Arab Emirates, 2019, pp. 1–6.
[38] S. Alshamasi, M. B. Menai, Ensemble-Based Clustering for Writing Style Change Detection in
Multi-Authored Textual Documents, in: CLEF 2022 Labs and Workshops, Notebook Papers, volume
3180 of CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022, p. 18.
[39] L. Mandic, F. Milkovic, S. Doria, Combining the Powers of Clustering Afinities in Style Change</p>
        <p>Detection, Course Project Reports, University of Zagreb, Zagreb, Croatia, 2019.
[40] C. Zuo, Y. Zhao, R. Banerjee, Style Change Detection with Feed-forward Neural Networks, in:
CLEF 2019 Labs and Workshops, Notebook Papers, volume 2380 of CEUR Workshop Proceedings,
CEUR-WS.org, Lugano, Switzerland, 2019, p. 9.
[41] N. Schaetti, Character-based Convolutional Neural Network for Style Change Detection, in:
CLEF 2018 Labs and Workshops, Notebook Papers, volume 2125 of CEUR Workshop Proceedings,
CEUR-WS.org, Avignon, France, 2018, p. 6.
[42] M. Hosseinia, A. Mukherjee, A Parallel Hierarchical Attention Network for Style Change Detection,
in: CLEF 2018 Labs and Workshops, Notebook Papers, volume 2125 of CEUR Workshop Proceedings,
CEUR-WS.org, Avignon, France, 2018, p. 7.
[43] X. Jiang, H. Qi, Z. Zhang, M. Huang, Style Change Detection: Method Based On Pre-trained Model
And Similarity Recognition, in: CLEF 2022 Labs and Workshops, Notebook Papers, volume 3180
of CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022, p. 6.
[44] Q. Lao, L. Ma, W. Yang, Z. Yang, D. Yuan, Z. Tan, L. Liang, Style Change Detection Based On
Bert And Conv1d, in: CLEF 2022 Labs and Workshops, Notebook Papers, volume 3180 of CEUR
Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022, p. 6.
[45] J. Zi, L. Zhou, Style Change Detection Based On Bi-LSTM And Bert, in: CLEF 2022 Labs and
Workshops, Notebook Papers, volume 3180 of CEUR Workshop Proceedings, CEUR-WS.org, Bologna,
Italy, 2022, p. 5.
[46] Z. Zhang, Z. Han, L. Kong, Style Change Detection based on Prompt, in: CLEF 2022 Labs and
Workshops, Notebook Papers, volume 3180 of CEUR Workshop Proceedings, CEUR-WS.org, Bologna,
Italy, 2022, p. 4.
[47] Z. Zhang, Z. Han, L. Kong, X. Miao, Z. Peng, J. Zeng, H. Cao, J. Zhang, Z. Xiao, X. Peng, Style
Change Detection Based On Writing Style Similarity, in: CLEF 2021 Labs and Workshops, Notebook
Papers, volume 2936 of CEUR Workshop Proceedings, CEUR-WS.org, Bucharest, Romania, 2021,
p. 4.
[48] T.-M. Lin, C.-Y. Chen, Y.-W. Tzeng, L.-H. Lee, Ensemble Pre-trained Transformer Models for
Writing Style Change Detection, in: CLEF 2022 Labs and Workshops, Notebook Papers, volume
3180 of CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022, p. 9.
[49] X. Liu, H. Chen, J. Lv, Team foshan-university-of-guangdong at PAN: Adaptive Entropy-Based
Stability-Plasticity for Multi-Author Writing Style Analysis, in: Working Notes of CLEF 2024
- Conference and Labs of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings,
CEUR-WS.org, Grenoble, France, 2024, pp. 2750–2754.
[50] T. M. Mohan, T. V. S. Sheela, BERT-Based Similarity Measures Oriented Approach for Style Change
Detection, in: Accelerating Discoveries in Data Science and Artificial Intelligence II, volume 438,
Springer Nature Switzerland, Cham, 2024, pp. 83–94.
[51] T.-M. Lin, Y.-H. Wu, L.-H. Lee, Team NYCU-NLP at PAN 2024: Integrating Transformers with
Similarity Adjustments for Multi-Author Writing Style Analysis, in: Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings,
CEUR-WS.org, Grenoble, France, 2024, pp. 2716–2721.
[52] Y. Huang, L. Kong, Team Text Understanding and Analysis at PAN: Utilizing BERT Series
Pretraining Model for Multi-Author Writing Style Analysis, in: Working Notes of CLEF 2024
Conference and Labs of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings,
CEUR-WS.org, Grenoble, France, 2024, pp. 2653–2657.
[53] Q. Wu, L. Kong, Z. Ye, Team bingezzzleep at PAN: A Writing Style Change Analysis Model Based
on RoBERTa Encoding and Contrastive Learning for Multi-Author Writing Style Analysis, in:
Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, volume 3740 of
CEUR Workshop Proceedings, CEUR-WS.org, Grenoble, France, 2024, pp. 2963–2968.
[54] Z. Chen, Y. Han, Y. Yi, Team Chen at PAN: Integrating R-Drop and Pre-trained Language Model
for Multi-author Writing Style Analysis, in: Working Notes of CLEF 2024 - Conference and Labs
of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings, CEUR-WS.org, Grenoble,
France, 2024, pp. 2547–2553.
[55] B. Wu, Y. Han, K. Yan, H. Qi, Team baker at PAN: Enhancing Writing Style Change Detection with
Virtual Softmax, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum,
volume 3740 of CEUR Workshop Proceedings, CEUR-WS.org, Grenoble, France, 2024, pp. 2951–2955.
[56] M. K. Sheykhlan, S. K. Abdoljabbar, M. N. Mahmoudabad, Team karami-sh at PAN:
Transformerbased Ensemble Learning for Multi-Author Writing Style Analysis, in: Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings,
CEUR-WS.org, Grenoble, France, 2024, pp. 2676–2681.
[57] C. Liu, Z. Han, H. Chen, Q. Hu, Team Liuc0757 at PAN: A Writing Style Embedding Method Based
on Contrastive Learning for Multi-Author Writing Style Analysis, in: Working Notes of CLEF
2024 - Conference and Labs of the Evaluation Forum, volume 3740 of CEUR Workshop Proceedings,
CEUR-WS.org, Grenoble, France, 2024, pp. 2716–2721.
[58] M. T. Zamir, M. A. Ayub, A. Gul, N. Ahmad, K. Ahmad, Stylometry Analysis of Multi-authored</p>
        <p>Documents for Authorship and Author Style Change Detection, 2024.
[59] M. Fröbe, M. Wiegmann, N. Kolyada, B. Grahm, T. Elstner, F. Loebe, M. Hagen, B. Stein, M. Potthast,
Continuous Integration for Reproducible Shared Tasks with TIRA.io, in: Advances in Information
Retrieval. 45th European Conference on IR Research (ECIR 2023), Lecture Notes in Computer
Science, Springer, Berlin Heidelberg New York, 2023, pp. 236–241.
[60] Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, J. M. Solomon, Dynamic Graph CNN for</p>
        <p>Learning on Point Clouds, ACM Transactions on Graphics 38 (2019) 1–12.
[61] A. S. Alsheddi, M. E. B. Menai, Edge Convolutional Networks for Style Change Detection in Arabic
Multi-Authored Text, Applied Sciences 15 (2025) 6633. URL: https://www.mdpi.com/2076-3417/15/
12/6633. doi:10.3390/app15126633.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Tschuggnall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Verhoeven</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Daelemans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Potthast, Overview of the Author Identi cation Task at PAN-2017: Style Breach Detection and Author Clustering</article-title>
          , in: Working Notes of CLEF 2017 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>1866</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Dublin, Ireland,
          <year>2017</year>
          , p.
          <fpage>22</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kestemont</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tschuggnall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Daelemans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Potthast, Overview of the Author Identi cation Task at PAN-2018</article-title>
          , in: Working Notes of CLEF 2018 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>2125</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Avignon, France,
          <year>2018</year>
          , p.
          <fpage>25</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tschuggnall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <source>Overview of the Style Change Detection Task at PAN</source>
          <year>2019</year>
          , in: Working Notes of CLEF 2019 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>2380</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Lugano, Switzerland,
          <year>2019</year>
          , p.
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          , G. Specht,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <source>Overview of the Style Change Detection Task at PAN</source>
          <year>2020</year>
          , in: Working Notes of CLEF 2020 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>2696</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Thessaloniki, Greece,
          <year>2020</year>
          , p.
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <source>Overview of the Style Change Detection Task at PAN</source>
          <year>2021</year>
          , in: Working Notes of CLEF 2021 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>2936</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Bucharest, Romania,
          <year>2021</year>
          , pp.
          <fpage>1760</fpage>
          -
          <lpage>1771</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zangerle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <source>Overview of the Style Change Detection Task at PAN</source>
          <year>2022</year>
          , in: Working Notes of CLEF 2022 -
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          , volume
          <volume>3180</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Bologna, Italy,
          <year>2022</year>
          , pp.
          <fpage>2344</fpage>
          -
          <lpage>2356</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Greiner-Petter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle, Overview of PAN 2025:
          <article-title>Voight-Kampf Generative AI Detection</article-title>
          , Multilingual Text
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>