<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>StarBERT: A Hybrid Neural Network Model for Human-AI Collaborative Text Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Miaoji Zheng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yong Zhong</string-name>
          <email>zhongyong@fosu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fen Liu</string-name>
          <email>liufen@fosu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tufeng Xian</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Meifang Xie</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Weidong Wu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiliang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Qiyuan Sun</string-name>
          <email>sunqiyuan@fosu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Foshan University</institution>
          ,
          <addr-line>Foshan</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>As large language models (LLMs) become more accessible, machine-generated content is rapidly increasing in many fields. These models can produce fluent and coherent text, making them useful for automating various writing tasks. However, their wide use has also raised concerns about misinformation, academic honesty, and the authenticity of content. Therefore, it is important to identify how much of a text is created by humans and how much by machines. In this study, we introduce StarBERT, a new hybrid model that combines DeBERTav3-large with StarBlock2d. This model focuses on classifying texts written through human-AI collaboration. Specifically, after dividing the texts into six categories based on the type of human and machine contribution, StarBERT uses the deep language understanding of DeBERTa-v3-large and the high-dimensional mapping ability of StarBlock2d. Our results show that StarBERT performs significantly better than existing baseline models, such as RoBERTa-base.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2025</kwd>
        <kwd>Human-AI Collaborative Text Classification</kwd>
        <kwd>BERT</kwd>
        <kwd>Star Operation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The rapid spread of generative AI models, such as GPT-4o and Claude 3.5, has ushered in a new era
of human-AI collaborative creation, where machine-generated text is deeply integrated with
humanauthored content. While this collaboration boosts productivity, it also presents serious challenges for
tracing the origin of texts and identifying authorship—both of which are essential for maintaining
academic integrity, combating misinformation, and ensuring content authenticity.</p>
      <p>
        Although some studies have ofered valuable insights [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], most existing research focuses on
specific application areas or remains limited to the traditional binary classification approach (human
vs. machine) in AI-generated text detection. For example, Liu et al. introduced the concepts of
graph representation and structural entropy to study model performance under imbalanced data
conditions, aiming to improve text classification and detection accuracy [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. To evaluate the robustness
of detectors on mixed-source generated texts, Huang G et al. proposed a new model called the Siamese
Calibration Reconstruction Network (SCRN) using the SeqXGPT-Bench dataset [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. SCRN introduces
and removes noise in the text through a reconstruction network, extracting semantic representations
that are robust to local perturbations, which aids in feature analysis. To further enhance detection
performance, Mo et al. developed an efective tool for detecting AI-generated text [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. This method
uses deep learning techniques by combining Transformer, Long Short-Term Memory (LSTM), and
Convolutional Neural Network (CNN) layers for eficient text classification and sequence labeling tasks.
In addition, the preprocessing steps are thorough, including Unicode normalization, case conversion,
removal of non-letter characters and extra spaces, and the use of specific delimiters for joining tokens.
Overall, this rigorous preprocessing pipeline ensures clean and consistent input for the model, and its
systematic approach and promising results highlight the tool’s potential for wide application and future
development in the field of AI-generated text detection (AIGTD). Further more, Wu et al. proposed
the BertT model to handle the Generative AI Authorship Verification task by leveraging BERT’s deep
semantic understanding capabilities and the eficient sequence processing power of Transformers [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Traditional AI text detection models that rely solely on binary classification (i.e., human vs. machine)
are not well-suited to handle the full spectrum of human-AI collaborative writing. These models
often fail to identify complex co-writing patterns, ranging from light AI-assisted editing to deeply
integrated joint narratives. To help people better understand human-AI collaboration and reduce the
risks associated with synthetic texts, PAN@CLEF 2025 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ](PAN is a series of scientific events and shared
tasks on digital text forensics and stylometry)introduced Subtask 2: Human-AI Collaborative Text
Classification in the Voight-Kampf Generative AI Detection task. This subtask focuses on classifying
documents co-created by humans and LLMs. To address these challenges, we propose the DeBERTa-Star
architecture, which deeply integrates the semantic understanding capabilities of DeBERTa-v3-large with
the high-dimensional feature transformation power of StarBlock2d. The core innovations of StarBERT
include:
• Utilizing the enhanced positional encoding of DeBERTa-v3-large [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] to accurately capture context
dependencies in collaborative texts.
• Applying a star operation (element-wise multiplication) to implicitly project the decoupled
attention outputs into a high-dimensional nonlinear space.
      </p>
      <p>StarBERT is trained using a stable cross-entropy loss function to enhance training robustness. It
was fine-tuned on a JSONL-formatted dataset provided by the organizers to better distinguish among
six types of human-AI collaboration. Following local training and initial validation, we submitted
the model’s predictions—formatted as "id": "identifier of the test sample", "label": 1—to the CodaLab
platform. This platform ofers a rigorous and controlled evaluation environment, ensuring fair and
transparent benchmarking against standard baselines. StarBERT achieved strong results across several
key metrics: Macro Recall of 57.46%, Macro F1 of 56.31%, and Accuracy of 66.81%, highlighting its
capability to distinguish human-AI collaborative writing efectively. These results demonstrate the
practical value of StarBERT in real-world collaborative text classification tasks.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Dataset Preprocessing</title>
      <p>In natural language processing tasks, data preprocessing is a crucial step. It not only lays the foundation
for model training but also directly impacts the model’s final performance. Our preprocessing pipeline
includes two main components: label conversion and text tokenization.</p>
      <p>We begin by converting the labels in the dataset. In the original data, labels are stored as strings,
while deep learning models typically require labels in integer format. To handle this, we define a
mapping function that converts the label field in each data sample to its corresponding integer value.
The label range spans from 0 to 5, representing six distinct categories. Next, we tokenize the text using
the tokenizer aligned with DeBERTa-v3-large. The tokenizer converts each text sample into a sequence
of token indices from the model’s vocabulary, while also indicating which tokens are valid and which
are padding. After tokenization, we apply truncation and padding to ensure that all input sequences
conform to the model’s maximum input length. We set this maximum to 512 tokens, so all inputs are
either truncated or padded to exactly 512 tokens. Once tokenization is complete, the data is ready
for model training. We ensure the quality of the input by carefully preprocessing the dataset, which
transforms raw data into a format suitable for deep learning models.</p>
      <p>This comprehensive preprocessing pipeline not only prepares the dataset for efective training of
StarBERT, but also enhances the model’s accuracy in distinguishing human-AI collaborative text—a key
requirement for this evaluation task.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>In this study, we introduce StarBERT, a hybrid neural network model that combines the powerful
feature extraction capabilities of DeBERTa-v3-large with the implicit high-dimensional feature fusion
mechanism of StarBlock2d to handle complex textual features and distinguish subtle diferences between
categories of human-AI collaborative writing. To help readers better understand StarBERT, we use
Figure 1 to illustrate its structure.</p>
      <p>We adopt the pretrained roberta-v3-large model from Hugging Face’s Transformers library as the
backbone BERT layer, leveraging its large-scale pretrained parameters for advanced natural language
understanding. This layer captures rich contextual information from complex inputs such as
JSONLformatted data and applies dropout between transitions to reduce overfitting and enhance robustness.</p>
      <p>
        On top of this, we integrate StarBlock, an eficient neural module centered on element-wise
multiplication (the Star Operation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]), which we adapt from the image domain to natural language processing
by proposing StarBlock2d. In our design, input features pass through two independent 1×1 2D
convolutional layers, one followed by a ReLU6 activation, and their outputs are fused via element-wise
multiplication to form an implicit high-dimensional representation that enables nonlinear feature
interactions similar to kernel methods. This result is then processed through a 3×3 depthwise separable
convolution (DWConv [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]) to extract spatial features, followed by batch normalization (BN) for training
stability. Overall, StarBlock2d achieves near-infinite expressive capacity within a low-dimensional
computational space, playing a key role in enabling StarBERT to outperform baselines and peer systems.
      </p>
      <p>
        During testing, StarBERT independently processes each text sample in JSONL format, evaluates its
likelihood of belonging to one of six categories—Fully human-written, Human-initiated then
machinecontinued, Human-written then machine-polished, Machine-written then humanized (obfuscated),
Machine-written then human-edited, or Deeply-mixed text—and classifies it based on the highest score.
By fine-tuning hyperparameters such as learning rate and batch size, and optimizing with binary
cross-entropy loss, the model is calibrated for high performance on key metrics such as Macro Recall
and Macro F1. This setup enables StarBERT to meet the specific challenges of the PAN@CLEF 2025
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] shared task on human-AI collaborative writing, demonstrating both theoretical innovation and
strong practical performance.
      </p>
      <p>text</p>
      <p>DeBERTa-v3- Dropout
large</p>
      <p>StarBlock2d</p>
      <p>Linear</p>
      <p>output six labels</p>
      <p>DW-Conv2d
FC</p>
      <p>GELU</p>
      <p>FC
sum</p>
      <p>StarBlock2d</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <sec id="sec-4-1">
        <title>4.1. Experiment Settings</title>
        <p>In our experimental setup, to evaluate the performance of StarBERT in the human-AI collaborative text
classification task, we used the oficially provided subtask2_train.jsonl as the training set and
subtask2_dev.jsonl as the validation set for training or fine-tuning. To ensure the reproducibility
of our experiments, we fixed the random seed to 42. This setting guarantees consistent results across all
stages of the training process, including data loading and model initialization. To balance computational
eficiency and learning depth, we trained the model on a CUDA-enabled GPU using the AdamW
optimizer. The maximum input length for the model was set to 512 tokens. The learning rate was set to
2e-5, which is a commonly efective value for most text classification tasks. The batch size was 8, and
the model was trained for one epoch.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Metrics</title>
        <p>In this task, three evaluation metrics were defined: Macro Recall, Macro F1, and Accuracy. These metrics
are designed to comprehensively reflect the model’s performance on the collaborative human-AI text
classification task. A detailed introduction to each metric is provided below.</p>
        <p>
          Macro Recall: Macro Recall is the arithmetic mean of the recall scores across all classes, giving equal
importance to the recognition ability of each class [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. It is defined as:
        </p>
        <p>Macro Recall = 1 ∑︁</p>
        <p>=1   +  
( = 6)
where   denotes the number of true positives for class  (correctly predicted samples), and  
represents the number of false negatives for class  (missed samples).</p>
        <p>
          Macro F1: Macro F1 is the arithmetic mean of the F1-scores across all classes, which balances both
precision and recall [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. The formula is defined as:
        </p>
        <p>Macro F1 =
 =1
1 ∑︁ 2 · Precision · Recall</p>
        <p>Precision + Recall</p>
        <p>This macro-averaging strategy calculates the F1-score for each class independently and then averages
them, which avoids bias toward majority classes.</p>
        <p>
          Accuracy: Accuracy measures the proportion of correctly predicted samples among all samples,
reflecting the overall classification performance [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. It is defined as:
∑︀=1
        </p>
        <p>total
Accuracy =</p>
        <p>(total = total number of samples)</p>
        <p>These three metrics jointly provide an efective and balanced evaluation of our model’s capability in
the human-AI collaborative text classification task.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Results</title>
        <p>Our StarBERT model demonstrated strong and stable performance in the PAN 2025 human-AI
collaborative text classification task, showing substantial efectiveness across several key evaluation metrics. As
shown in Table 1, StarBERT achieved a Macro Recall of 57.46%, significantly outperforming the baseline
model’s 48.32%. This reflects its strong ability to accurately identify instances across all categories.
Moreover, StarBERT achieved a Macro F1 score of 56.31%, well above the baseline’s 47.82%, indicating
(1)
(2)
(3)
that the model efectively balances predictive performance across categories. It avoids bias toward
classes with more samples, which often occurs in imbalanced datasets. In terms of Accuracy, StarBERT
reached 66.81%, again far exceeding the baseline score of 57.09%, ofering a clear and intuitive measure
of the model’s overall classification capability in this task. In the oficial PAN 2025 leaderboard, our
submission ranked 4th out of 22 participating teams. This result underscores the competitive advantage
of our approach in efectively capturing discriminative textual features in the context of human-AI
collaboration.</p>
        <p>These outcomes validate that StarBERT not only embodies theoretical innovation but also
demonstrates substantial practical efectiveness in this domain. The model is capable of distinguishing
collaborative human-AI texts with high accuracy, making it a valuable tool for complex text analysis
tasks.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>StarBERT integrates the deep semantic understanding capabilities of DeBERTa-v3-large with the
high-dimensional feature mapping power of StarBlock2d, achieving—for the first time—precise
sixclass classification of human-AI collaborative texts. Experimental results demonstrate that the model
significantly outperforms baseline models (e.g., RoBERTa-base) in the PAN 2025 task, achieving a
Macro Recall of 57.46%, a Macro F1 score of 56.31%, and an Accuracy of 66.81%. These results validate
StarBERT’s superiority in handling mixed-authorship texts.</p>
      <p>The innovative Star Operation and enhanced positional encoding efectively capture the nonlinear
characteristics and contextual dependencies inherent in collaborative texts. This research provides a
scalable technical solution for promoting academic integrity and supporting content provenance in the
era of AI-assisted writing.</p>
      <p>Future work will focus on further optimizing model parameters, enhancing feature engineering
techniques, and expanding the diversity of the training dataset. These improvements aim to increase
the model’s generalization ability and performance across various textual contexts. This continued
development is expected to further improve StarBERT’s detection accuracy and extend its applicability
in real-world scenarios.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgements</title>
      <p>This work was supported by grants from the Guangdong-Foshan Joint Fund Project (No.
2022A1515140096) and Open Fund for Key Laboratory of Food Intelligent Manufacturing in Guangdong
Province (No. GPKLIFM-KF-202305).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used GPT-4 and DeepSeek-R1 in order to: Grammar
and spelling check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as
needed. full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Ghosal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geiping</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Manocha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Bedi</surname>
          </string-name>
          ,
          <article-title>Towards possibilities &amp; impossibilities of ai-generated text detection: A survey</article-title>
          ,
          <source>arXiv preprint arXiv:2310.15264</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Ma, J. Liu,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <article-title>Detection vs. anti-detection: Is text generated by ai detectable?</article-title>
          , in: International Conference on Information, Springer,
          <year>2024</year>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>222</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <article-title>Coco: Coherence-enhanced machine-generated text detection under data limitation with contrastive learning</article-title>
          ,
          <source>arXiv preprint arXiv:2212.10341</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>You</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Are ai-generated text detectors robust to adversarial perturbations?</article-title>
          ,
          <source>arXiv preprint arXiv:2406.01179</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Large language model (llm) ai text generation detection based on transformer deep learning algorithm</article-title>
          ,
          <source>arXiv preprint arXiv:2405.06652</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Bertt: a hybrid neural network model for generative ai authorship verification</article-title>
          , Working Notes of CLEF (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tsivgun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abassy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mansurov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Ta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Elozeiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. V.</given-names>
            <surname>Tomar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Artemova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Habash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the “VoightKampf” Generative AI Authorship Verification Task at PAN</article-title>
          and
          <article-title>ELOQUENT 2025</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , W. Chen,
          <article-title>Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing</article-title>
          ,
          <source>arXiv preprint arXiv:2111.09543</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <article-title>Rewrite the stars</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>5694</fpage>
          -
          <lpage>5703</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          , Xception:
          <article-title>Deep learning with depthwise separable convolutions</article-title>
          ,
          <source>in: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1251</fpage>
          -
          <lpage>1258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Greiner-Petter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle, Overview of PAN 2025:
          <article-title>Generative AI Authorship Verification, Multi-Author Writing Style Analysis, Multilingual Text Detoxification, and Generative Plagiarism Detection, in: Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sokolova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lapalme</surname>
          </string-name>
          ,
          <article-title>A systematic analysis of performance measures for classification tasks</article-title>
          ,
          <source>Information processing &amp; management 45</source>
          (
          <year>2009</year>
          )
          <fpage>427</fpage>
          -
          <lpage>437</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>C. M. Bishop</surname>
            ,
            <given-names>N. M.</given-names>
          </string-name>
          <string-name>
            <surname>Nasrabadi</surname>
          </string-name>
          ,
          <article-title>Pattern recognition and machine learning</article-title>
          , volume
          <volume>4</volume>
          , Springer,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. L. Thompson</surname>
          </string-name>
          , W. Zucchini,
          <article-title>On the statistical analysis of roc curves</article-title>
          ,
          <source>Statistics in medicine 8</source>
          (
          <year>1989</year>
          )
          <fpage>1277</fpage>
          -
          <lpage>1290</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>