<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>DeBERTa-FPN: Fusion Feature Pyramid Network for Human-AI Collaborative Text Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Qiyuan Sun</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Li Ma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wenyin Yang</string-name>
          <email>cswyyang@fosu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tufeng Xian</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Meifang Xie</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Weidong Wu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiliang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Miaoji Zheng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Foshan University</institution>
          ,
          <addr-line>Foshan</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>In recent years, large language models (LLMs) have developed rapidly and have been widely used in the field of text generation. The increasingly realistic generated content has led to some security risks in information dissemination. Human-machine collaborative text classification has become a critical task and is extremely challenging. This paper proposes a human-machine collaborative text classification model, namely DeBERTa-FPN, which combines DeBERTa-V3 and feature fusion pyramid (FPN), aiming to use the powerful text processing capabilities of DeBERTa-V3-Large and the multi-scale feature extraction capabilities of FPN to improve the model performance of human-machine collaborative text classification. The introduction of FPN can enhance the model's function in feature extraction, and also more efectively combine global features to help complete the classification task. Experimental results show that our method significantly outperforms the baseline model, with 12% increase in Recall, 13% increase in F1, and 10% increase in Accuracy. Thus, we have verified the efectiveness of this method.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2025</kwd>
        <kwd>Human-AI Collaborative Text Classification</kwd>
        <kwd>DeBERTa-V3</kwd>
        <kwd>Feature Fusion Pyramid</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The goal of the human-machine collaborative text classification task in PAN@CLEF2025 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is to
classify text into six categories based on the nature of human and machine contributions: fully
humanwritten, human-initiated and then machine-continued, human-abbreviated and then machine-polished,
machine-written and then machine-humanized, machine-written and then human-edited, and deep hybrid
text [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. As the quality of machine-generated text is getting better and better, and it is comparable to human
writing in terms of logic and vividness, existing classification methods usually do not work well in this task,
and more efective methods are needed to complete the classification task.
      </p>
      <p>In order to solve the problem of text writing classification and distinguish diferent writing types, this
paper proposes DeBERTa-FPN, a new method that combines the DeBERTa-V3-Large pre-trained model and
feature fusion pyramid to meet these challenges. As a pre-trained model, DeBERTa-V3-Large has strong
context understanding ability and can capture detailed features in text. Feature Fusion Pyramid (FPN) is a
multi-scale feature fusion module. This method was originally proposed in image classification tasks to link
global and local features of images to improve classification eficiency. In this task, we use its multi-scale
characteristics to associate features of texts of diferent lengths. By combining text features of diferent
lengths, we efectively utilize global and local feature information, enhance the model’s perception of
features, and ensure eficient and accurate classification. By combining the two, our model can efectively
complete the classification task in the human-machine collaborative text classification task.</p>
      <p>To evaluate the efectiveness of DeBERTa-FPN, we tested it through the submission platform
speciifed by PAN. The platform provides a strictly controlled testing environment to ensure fair and transparent
benchmarking based on established benchmarks. This initial submission is crucial to evaluating the
practicality of the model in real-world scenarios and improving its performance based on objective feedback.
After this evaluation, DeBERTa-FPN performed well in multiple key indicators, with the Recall index of
54.49%, F1 index of 54.4%, and Accuracy of 62.89%. These results are significantly better than baseline
models such as RoBERTa-base and DetectGPT, which shows the efect of DeBERTa-FPN in distinguishing
human-machine collaborative text and the efectiveness of our method in solving human-machine
collaborative text classification tasks.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        In recent years, text generation artificial intelligence tools have developed rapidly, and the generation
efect has been significantly improved in logic and vividness. As the quality of text generated by LLMs
(such as GPT-4, BERT, etc.) is close to human level, its wide application in education, medical care, news,
law and other fields has also brought ethical issues, such as the spread of false information, academic
plagiarism and public opinion manipulation. Therefore, it is urgent to develop eficient AI-generated text
detection technology[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. For the task of human-computer collaborative text classification, early
humancomputer text classification research mainly focused on binary classification tasks, that is, distinguishing
whether the text is written by humans or generated by AI. Jinyan Su et al. proposed DetectLLM-LRR and
DetectLLM-NPR for detecting machine-generated text[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. LRR measures text features by the ratio of
loglikelihood to log-rank, while NPR generates perturbed text by making small perturbations to the original
text and then calculates the normalized log-rank mean of these perturbed texts. Both methods show good
recognition results and can provide efective feedback in a short time.However, with the popularization of
AI writing assistance tools, researchers gradually realized that this simple binary division can no longer
meet actual needs, and gradually turned to the study of multi-class division[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. In order to improve the
classification efect, the Mindner team systematically evaluated 37 features, covering 8 categories
including perplexity, semantics, list search, documents, error-based, readability, AI feedback and text vectors,
all of which can be used as the basis for distinguishing human-machine writing[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Schaaf et al. used
machine learning methods for multilingual AI-generated text detection and obtained good classification
performance through multi-layer perceptron (MLP), but the detection performance for AI rewriting was
poor[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In addition to traditional machine learning methods, large models have also been used in
humanmachine collaborative text classification. Kumar et al. used the BERT model to obtain text embeddings
to compare the text similarity between people and between people and AI[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Sun et al. explored how to
more efectively utilize the results of BERT model preprocessing[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In more extensive research, it was
found that pre-trained models such as RoBERTa[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and DeBERTa-V3[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] also played an important role
in text classification tasks. They can be used as modules for text feature extraction to provide more
efective text feature information[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. However, pre-trained models often fail to focus on global information
during feature extraction, and only focus on deep features, which may cause some useful features to be
ignored. Feature fusion pyramid was first proposed for target detection tasks to fuse high-level semantic
information with underlying detail features, significantly improving the detection accuracy of small
targets[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. This method has now been widely used in many fields to fuse multiple features to improve the
efect of feature extraction. This study aims to enhance the perception ability of the pre-trained model
for high-level and low-level features by using the feature fusion pyramid method, so as to achieve a more
efective human-computer collaborative text classification model.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. System Overview</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset and Preprocessing</title>
        <p>This study used the “Human-machine Collaborative Text Classification Task” dataset provided by
PAN@CLEF. This dataset is a publicly available dataset specifically designed to verify human-machine
collaborative text classification. The PAN@CLEF dataset contains Multi-domain documents such as academic,
news, and social media, Human-written and machine-generated samples, Collaborative texts with
annotation layers for human/machine contributions, and these texts are in multiple languages. The PAN@CLEF
dataset is usually organized into the following format, which contains text content, language type, label,
data source, model type, and label type information:
{"text":"...","language":"...","label":0,"source_dataset":"...","model":"...","label_text":"fully
human-written"}
{"text":"...","language":"...","label":1,"source_dataset":"...","model":"...","label_text":"human-written,
then machine-polished"}</p>
        <p>The validation set provided by the ”Human-Computer Collaborative Text Classification Task” in
PAN@CLEF is a key component for testing and optimizing the author’s validation model. Its
organization is consistent with the above. It should be mentioned that the label and label type are linked in the
dataset, and 0-5 is used to represent diferent text types when predicting. Its comparison is as follows:
{ 0: "fully human-written",
1: "human-written, then machine-polished",
2: "machine-written, then machine-humanized",
3: "human-initiated, then machine-continued",
4: "deeply-mixed text (human + machine parts)",
5: "machine-written, then human-edited"}</p>
        <p>The sample sizes of the training set and development set are 288,918 and 72,661 respectively. The
specific number of categories is shown in Table 1. This study uses the PAN@CLEF dataset to train and
evaluate a hybrid model that combines DeBERTa-V3-Large and FPN, aiming to improve the accuracy of
human-machine collaborative text classification.</p>
        <p>Specifically, this dataset helps us systematically understand and identify the characteristics of diferent
types of human-machine collaborative text. Through a series of experiments on the PAN@CLEF dataset,
we evaluate the performance of the model in distinguishing six types of human-machine collaborative text
classification. We use multiple evaluation metrics such as precision, recall, and F1 score to
comprehensively analyze the efectiveness of the model. The results show that our model can achieve satisfactory
performance when dealing with this specific task.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Network Architecture</title>
        <p>In our study, we designed and implemented a hybrid neural network model named DeBERTa-FPN
that combines DeBERTa-V3-Large and FPN to perform complex human-machine collaborative text
classiifcation, aiming to distinguish six types of text with diferent machine participation. The structure of the
model is shown in Figure 1. The model architecture aims to make full use of the deep semantic processing
capabilities of DeBERTa-V3-Large and the global information integration and extraction capabilities of the
FPN module to enhance the model’s performance in handling fine-grained text analysis tasks. We use the
pre-trained DeBERTa-V3-Large model provided by Hugging Face as the pre-processing module for text
information. This pre-trained model has a 24-layer Transformer, which can process long texts, more
accurately model inter-word dependencies, capture long-distance dependencies in text, and obtain efective
features for classification. In order to better combine global and detailed features, we use FPN as a feature
enhancement module. FPN can fuse feature information at diferent levels in DeBERTa-V3-Large, enhance
the model’s ability to understand global information, improve the utilization of features, and provide more
efective features for the classifier. For our model, we used the features of the 12th and 18th layers for
fusion. These two levels are chosen because they not only ensure the adequacy of feature extraction, but
also combine feature information at diferent levels to provide efective information for further fusion.
They were processed by Conv1d for deep feature extraction and concatenated before being passed to the
Max Pooling layer to enhance key features. Finally, the features were passed to the classifier, where a
dropout layer was added to prevent overfitting. Finally, the features were mapped to the classification
results through a fully connected layer to achieve classification of six types of human-machine collaborative
texts.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <sec id="sec-4-1">
        <title>4.1. Experiment Settings</title>
        <p>In this study, we used the pre-trained DeBERTa-V3-Large model as the basis for text feature extraction.
Its large resource consumption also provides very impressive text feature extraction capabilities, so it is
necessary to select the most efective training scheme in combination with time and computing resources.
During the training process, we selected a batch size of 2, a learning rate of 1e-5, and a total of 3 training
cycles to ensure that the model can efectively learn text features and avoid overfitting. In addition, we
used the AdamW optimizer, which was selected for its optimization efect in deep learning model training,
especially in dealing with gradient sparsity and weight decay. To ensure the repeatability of the experiment,
we set a fixed random seed, and all experiments were conducted in a computing environment equipped
with NVIDIA GeForce RTX 3090 and Intel(R) Core(TM) i9-10900K.</p>
        <p>The main experimental process includes three stages: data preparation, model training, and
performance evaluation. First, in the data preparation stage, the dataset is preprocessed, including text cleaning,
label correspondence, etc. In addition, the dataset is divided into a training set and a validation set. During
the model training phase, the model is iteratively learned on the training set. At the end of each cycle, we
evaluate the performance of the model on the validation set to monitor whether there is overfitting during
the training process. Finally, in the performance evaluation phase, we use standard classification metrics
such as accuracy, recall, and F1 to evaluate the model. We pay special attention to the performance of the
model on an independent test set to verify its generalization ability in practical applications. Through this
series of meticulous and rigorous experimental processes, we ensure the accuracy and practicality of the
research results.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results</title>
        <p>In order to comprehensively evaluate the performance of our proposed DeBERTa-FPN, we selected a
series of indicators, including F1, recall, and Accuracy. These indicators not only reflect the overall
performance of the model, but also provide diferent performance evaluation perspectives to help us understand
the performance of the model in specific aspects.</p>
        <p>The specific performance metrics are as follows:</p>
        <p>Accuracy represents the proportion of samples correctly predicted by the model to the total samples,
which can intuitively reflect the overall prediction accuracy. It is calculated as:</p>
        <p>Accuracy =</p>
        <p>+  
  +   +   +  
(1)</p>
        <p>Among them, TP stands for True Positive, that is, the number of correctly judged positive examples,
TN stands for True Negative, that is, the number of correctly judged negative examples, FN stands for False
Negative, that is, the number of positive examples that are not judged correctly, and FP stands for False
Positive, that is, the number of negative examples that are mistaken for positive examples.</p>
        <p>Recall represents the proportion of samples that are actually positive and are correctly predicted. It
measures the ability of the model to predict correctly. It is calculated as:</p>
        <p>F1 score is the harmonic mean of precision and recall.It is calculated as:</p>
        <p>Recall =</p>
        <p>+</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Where</p>
      <p>Precision =   (4)</p>
      <p>+</p>
      <p>Through these evaluation indicators, we can learn to evaluate the overall performance of the model
and conduct in-depth analysis of the specific capabilities of the model from multiple dimensions. These
evaluation indicators can also objectively measure the performance diference between our proposed model
and other models, ensure the comprehensiveness and reliability of the evaluation results, and lay a solid
foundation for future model optimization and application. Through the above indicators, we compare with
the baseline model. Compared with the baseline model, our model achieved 54.49%, 54.40%, and 62.89% in
Recall, F1 Score, and Accuracy, respectively, which are 12%, 13%, and 10% higher than the baseline model.
The evaluation results under the evaluation set are shown in Table 2.</p>
      <p>Approach
Baseline
DeBERTa-FPN</p>
      <p>Accuracy (%) F1 Score (%) Recall (%)</p>
      <p>In this study, we proposed and implemented a hybrid neural network model that combines
DeBERTaV3-Large and FPN, aiming to improve the accuracy of human-machine collaborative text classification.
The results show that the Deberta-FPN hybrid model proposed in this study has achieved satisfactory
results in text classification tasks. It has achieved impressive results in multiple indicators. Compared
with the baseline model, the Recall indicator is improved by 12%, the F1 indicator is improved by 13%, and
the Accuracy is improved by 10%, which shows the efectiveness of the model structure and can achieve
excellent performance in related tasks. At the same time, we also know that there are still many aspects
of this model that can be improved. In future work, we will continue to improve the method and strive to
obtain better results in the task of human-machine collaborative text classification.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgements</title>
      <p>This work was supported by grants from the Guangdong-Foshan Joint Fund Project (No.
2022A1515140096) and Open Fund for Key Laboratory of Food Intelligent Manufacturing in Guangdong
Province (No. GPKLIFM-KF-202305).</p>
    </sec>
    <sec id="sec-7">
      <title>7. Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used DeepSeek-R1 in order to: Grammar and spelling
check. After using this tool, the authors reviewed and edited the content as needed. full responsibility for
the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <surname>AndréGreiner-Petter</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Karlgren</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Mayerl</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Panchenko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Shelmanov</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle, Overview of PAN 2025:
          <article-title>Voight-Kampf Generative AI Detection, Multilingual Text Detoxification, Multi-Author Writing Style Analysis, and Generative Plagiarism Detection</article-title>
          , in: J.
          <string-name>
            <surname>C. de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tsivgun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abassy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mansurov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Ta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Elozeiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. V.</given-names>
            <surname>Tomar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Artemova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Habash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the “Voight-Kampf” Generative AI Authorship Verification Task at PAN</article-title>
          and
          <article-title>ELOQUENT 2025</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fariello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Fenza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Forte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gallo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marotta</surname>
          </string-name>
          ,
          <article-title>Distinguishing human from machine: A review of advances and challenges in ai-generated text detection</article-title>
          ,
          <source>International Journal of Interactive Multimedia and Artificial Intelligence</source>
          <volume>8</volume>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhuo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          , Detectllm:
          <article-title>Leveraging log rank information for zero-shot detection of machine-generated text</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <year>2023</year>
          , pp.
          <fpage>12395</fpage>
          -
          <lpage>12412</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Abburi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Bowen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pudota</surname>
          </string-name>
          ,
          <article-title>Ai-generated text detection: A multifaceted approach to binary and multiclass classification</article-title>
          ,
          <source>arXiv preprint arXiv:2505.11550</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Mindner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Schlippe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Schaaf</surname>
          </string-name>
          ,
          <article-title>Classification of human-and ai-generated texts: Investigating features for chatgpt</article-title>
          ,
          <source>in: International conference on artificial intelligence in education technology</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>152</fpage>
          -
          <lpage>170</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Schaaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Schlippe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Mindner</surname>
          </string-name>
          ,
          <article-title>Classification of human-and ai-generated texts for english, french, german, and spanish</article-title>
          ,
          <source>arXiv preprint arXiv:2312.04882</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Prasad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Arti</surname>
          </string-name>
          ,
          <article-title>Comparative analysis of human and ai generated text</article-title>
          ,
          <source>in: 2024 11th International Conference on Signal Processing and Integrated Networks (SPIN)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>168</fpage>
          -
          <lpage>173</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yang</surname>
          </string-name>
          , L. Ma,
          <article-title>Bcav: a generative ai author verification model based on the integration of bert and cnn</article-title>
          , Working Notes of CLEF (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , W. Chen,
          <article-title>Debertav3: Improving deberta using electra-style pre-training with gradientdisentangled embedding sharing</article-title>
          ,
          <source>arXiv preprint arXiv:2111.09543</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zeng</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gašević</surname>
          </string-name>
          , G. Chen,
          <article-title>Detecting ai-generated sentences in human-ai collaborative hybrid texts: Challenges, strategies, and insights</article-title>
          ,
          <source>arXiv preprint arXiv:2403.03506</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>T.-Y. Lin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dollár</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Girshick</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Hariharan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Belongie</surname>
          </string-name>
          ,
          <article-title>Feature pyramid networks for object detection</article-title>
          ,
          <source>in: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>2117</fpage>
          -
          <lpage>2125</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>