<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Team Advacheck at PAN: Multitasking Does All the Magic</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anastasia Voznyuk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>German Gritsai</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrey Grabovoy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Advacheck OÜ</institution>
          ,
          <addr-line>Tallinn</addr-line>
          ,
          <country country="EE">Estonia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Université Grenoble Alpes (UGA)</institution>
          ,
          <addr-line>Grenoble</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The paper describes a system designed by Advacheck team to recognise sophisticated compound machinegenerated and human-written texts in two subtasks at the Voight-Kampf Generative AI Detection 2025 workshop organised as part of PAN 2025. Our developed system is a multi-task architecture with shared Transformer Encoder between several classification heads. As multiclass heads were trained to distinguish the domains presented in the data, they provide a better understanding of the samples. This approach led us to achieve 99.95% mean metric on validation set in Task 1 and the third place in the oficial ranking in Task 2 with 60.85% macro 1-score on the test set and bypass the baseline by 13%.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2025</kwd>
        <kwd>Voight-Kampf Generative AI Detection</kwd>
        <kwd>Human-AI Collaborative Text Classification</kwd>
        <kwd>natural language processing</kwd>
        <kwd>large language models</kwd>
        <kwd>multi-task learning</kwd>
        <kwd>domain adaptation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        As large language models are now firmly embedded in our world, helping researchers and internet
users alike, we increasingly interact with machine-generated text. However, as more and more of the
web becomes filled with generated content, the need for reliable detection methods grows. Potential
misuse includes malicious applications by students [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] or scientists [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4, 5</xref>
        ]. The mentioned things are
encouraging researchers to improve methods for detecting artificial text simultaneously with enhancing
generation methods.
      </p>
      <p>
        The task of detecting machine-generated texts is usually formulated as a binary text classification
task [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The most common solutions are to fine-tune the Transformer-based model [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] or to use
zero-shot approaches with intrinsic statistics of the text [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. While these methods perform well on
in-domain tasks [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], they are not robust to change of the domain, generator model, or language of
the texts [
        <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
        ]. Meanwhile, for the detection of AI-content in the wild such a change is, on the
contrary, a more realistic setup. Additionally, the quality of available data can be low, introducing noise
and making the detection task even harder [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The goal, then, is to build a model that can handle
noisy inputs and generalize to new domains.
      </p>
      <p>
        Because of the high coherence and fluency of modern LLMs, it is hard to find a simple, clear-cut
feature that separates human-written from machine-generated text. One promising direction is to
improve the representation space using multi-task learning (MTL) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. MTL has also shown strong
results in past competitions [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ], which encouraged us to use it in our approach. In this paper we
discuss our solution as the Advacheck team at Voight-Kampf Generative AI Detection 2025 [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Our
method shows that with additional internal data analysis and embedding alignment using MTL, it is
still possible to achieve high performance in detecting fragments in cross-domain and cross-generator
setups on texts from the advanced LLMs. As we forced model to focus on various domains, it allowed
us to form a cluster domain-wise structure for the text representations in the vector space. Additionally,
we applied model to a setup where one needs to classify diferent types of human-AI mixed writing
which can also be seen as “domains”. In our research, we show that (1) multi-task learning outperforms
the default single-task, (2) multi-task can perform well even for non-obvious domain setup such as
diferent ways of mixed human-AI writing.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Tasks Definition</title>
      <p>
        2.1. Task 1
The first subtask, Voight-Kampf AI Detection Sensitivity is binary AI detection task in that participants
are given a text and have to decide whether it was machine-authored (class 1) or human-authored (class
0). The organizers presented new models and diferent ways of obfuscating texts using LLM tools such
as style or mimicry. In addition, the data were augmented with new texts generated by participants
in the ELOQUENT task [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. The statistics of the dataset are summarised in Appendix A. The oficial
evaluation metrics are the following:
• ROC-AUC,
• Brier, the complement of the Brier score (mean squared loss),
• C@1: A modified accuracy score that assigns non-answers (score = 0.5) the average accuracy of
the remaining cases,
• 1 measure,
• 0.5 : a modified  0.5 measure (precision-weighted F-measure) that treats non-answers (score
= 0.5) as false negatives,
• the arithmetic mean of all the metrics above.
2.2. Task 2
The second subtask, Human-AI Collaborative Text Classification is stated as multiclassification task,
however, the classes are novel:
• Fully human-written: the document is entirely authored by a human without any AI assistance;
• Human-initiated, then machine-continued: a human starts writing, and an AI model
completes the text;
• Human-written, then machine-polished: the text is initially written by a human but later
refined or edited by an AI model;
• Machine-written, then machine-humanized (obfuscated): an AI generates the text, which is
later modified to obscure its machine origin;
• Machine-written, then human-edited: the content is generated by an AI but subsequently
edited or refined by a human;
• Deeply-mixed text: the document contains interwoven sections written by both humans and
      </p>
      <p>AI, without a clear separation;</p>
      <p>The oficial metric is Macro Recall, with additional metrics of accuracy and Macro 1.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Multitask Learning</title>
      <p>
        Similar to the previous competition on detection, COLING 2025, training set contained dozens of
generators and several distinct domains. Moreover, organisers claimed there will be more domains and
obfuscations in the test set. In the training data, we observed a large amount of valuable information
such as the name of a particular model, genres, which can sharpen the representations of each text
in the latent space and regularise the model. Therefore, following the best performed approach from
COLING [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], we decided to utilise multitask learning once again, as multitask learning may help the
model focus on those features that actually matter by shaping representations from all subtasks. Our
ifnal goal was to obtain fine-grained representations of the data that would ignore data-dependent
noise and generalises well. Since diferent tasks involve distinct noises, a model trained on multiple
tasks simultaneously is able to learn a more general representation. Furthermore, it reduces the risk of
overfitting.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Model for Task 1</title>
        <p>We propose an MTL architecture with hard parameter sharing (HPS), it is depicted in Figure 1. In HPS, a
common Transformer-based encoder is used for multiple tasks. After several variations of set of parallel
heads, we focused on three custom classification heads (CCH) for simultaneous training:
• Binary CCH for solving the initial task [2 classes]
• Multiclass CCH to define a genre [3 classes]
• Multiclass CCH determining model family belonging [4 classes]</p>
        <p>
          In this setup, the data with genres for the second head were taken from the original data proposed
by the authors, while the data for the third head passed a slight preprocessing. If we look at the list
of models proposed by the organisers, we can cluster most of them into families. We performed such
clustering by selected families of models, detailed description in Appendix A. These families did not
include rows with all the models represented in the original dataset, due to the small amount in some of
them. The idea of families should help to bring new insights into the representations, but not confuse
the model too much, as would be the case if we set this head to the task of classifying into all available
models without separating into families. Experiments with this sort of thing have been done as part of
our previous work [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. We chose DeBERTa-v3 base for the baseline and the backbone in our system,
as it is currently state-of-the-art model for supervised fine-tuning for binary classification [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>
          The model was trained in two phases: 1) fine-tuning assigned classifiers with frozen shared encoder
weights and 2) fine-tuning the complete model with all weights unfrozen . Therefore, at the first stage,
only the weights of the classifiers are updated, while at the second stage, all the weights in the model
are updated. These learning stages help to shift the distribution of the encoder weights in the right
direction and avoid overfitting [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. At the inference stage, only encoder with binary CCH predictions
used for final classification.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Model for Task 2</title>
        <p>Model for Task 2 was trained in the same way. We hypothesized that what really matters is who is the
initial author of the text and clustered the data based on this. After that we came up with the following
architecture:
• Multiclass CCH head for solving the initial task [6 classes];
• Multiclass CCH head for machine-based writing for classes Machine-written, then
machinehumanized and Machine-written, then human-edited into 3 domains related to the datasets [3
classes];
• Multiclass CCH head for human-based writing for classes fully human-written and human-written,
then machine-polished into 4 domains related to the datasets [4 classes];
• Multiclass CCH head for mixed writing for classes deeply-mixed text and human-initiated, then
machine-continued [6 classes];</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In Table 1 we show the preliminary results obtained on the validation set for the Task 1, together with
oficial baselines. In Table 2 we demonstrate the final results on the test set for the Task 2. Our approach
took 3rd place and performed much better than the suggested baseline.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper we described the system by the Advacheck team for both tasks at Voight-Kampf Generative
AI Detection 2025. We proposed solution with multi-task learning architecture that consists of shared
Transformer Encoder and composition of one binary and two multiclass Custom Classification Heads.
This approach led us to achieve 99.95% mean metric on validation set in Task 1 and outperformed all
the baselines on the test set. The described multi-task way of learning also allowed us to take the third
place in the oficial ranking in Task 2 with 60.85% macro 1-score on the test set and bypass the baseline
by 13%. Adding tasks for training in parallel reveal the formation of a cluster structure in the space of
embeddings, helping to achieve high results despite the presence of a large amount of noisy data.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT and Writefull in order to check grammar
and spelling and to paraphrase some sentences to improve clarity. After using this tool/service, the
author(s) reviewed and edited the content as needed and take(s) full responsibility for the publication’s
content.</p>
    </sec>
    <sec id="sec-7">
      <title>A. Data description</title>
      <p>This appendix reports the statistics within the provided training dataset for Subtask 1. Statistics are
presented for the target task of binary classification, genres and author attribution. More detailed in
Figure 2.</p>
      <p>In the shape of the architectural features that we came up with, we needed to cluster the models that
generated the examples. We were able to do this in 4 classes, some of the models were not included in
this data because their occurrences were not numerous and adding them to one of the existing groups
could confuse the model. In the merging we maintained the following families and their components:
• GPT family: gpt-3.5-turbo, gpt-4o-mini, gpt-4o, o3-mini, gpt-4.5-preview, gpt-4-turbo-paraphrase,
gpt-4-turbo
• Mistral family: ministral-8b-instruct-2410, mistral-7b-instruct-v0.2, mixtral-8x7b-instruct-v0.1
• Gemini family: gemini-2.0-flash, gemini-1.5-pro, gemini-pro, gemini-pro-paraphrase
• LLaMA family: llama-3.1-8b-instruct, llama-3.3-70b-instruct, llama-2-70b-chat, llama-2-7b-chat</p>
    </sec>
    <sec id="sec-8">
      <title>B. Custom Classification Head</title>
      <p>
        In our approach, we replaced the default one-layer linear classifier with a more extended version by
adding multiple layers, the final structure of Custom Classification Head (CCH) is shown in Figure
3. We chose GELU [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] as the activation feature and added dropout. In earlier experiments, when
compared with the base head, this adaptation gives a higher quality therefore we used it in all subsequent
experiments.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gašević</surname>
          </string-name>
          , G. Chen,
          <article-title>Towards automatic boundary detection for human-ai collaborative hybrid essay in education, 2023</article-title>
          . arXiv:
          <volume>2307</volume>
          .
          <fpage>12267</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Koike</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kaneko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Okazaki</surname>
          </string-name>
          , Outfox:
          <article-title>Llm-generated essay detection through in-context learning with adversarially generated examples</article-title>
          ,
          <source>in: AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ma</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yi</surname>
          </string-name>
          , Q. Cheng, Y. Huang,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          , X. Liu,
          <article-title>Ai vs</article-title>
          . human
          <article-title>- diferentiation analysis of scientific content generation</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2301</volume>
          .
          <fpage>10416</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Gritsay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabovoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kildyakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chekhovich</surname>
          </string-name>
          ,
          <article-title>Artificially generated text fragments search in academic documents</article-title>
          , in: Doklady Mathematics, volume
          <volume>108</volume>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>S434</fpage>
          -
          <lpage>S442</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Avetisyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Gritsay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabovoy</surname>
          </string-name>
          ,
          <article-title>Cross-lingual plagiarism detection: Two are better than one</article-title>
          ,
          <source>Program. Comput. Softw</source>
          .
          <volume>49</volume>
          (
          <year>2023</year>
          )
          <fpage>346</fpage>
          -
          <lpage>354</lpage>
          . doi:
          <volume>10</volume>
          .1134/S0361768823040138.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jawahar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdul-Mageed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lakshmanan</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.S.</surname>
          </string-name>
          ,
          <article-title>Automatic detection of machine generated text: A critical survey</article-title>
          , in: D.
          <string-name>
            <surname>Scott</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Bel</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          Zong (Eds.),
          <source>Proceedings of the 28th International Conference on Computational Linguistics</source>
          ,
          <source>International Committee on Computational Linguistics</source>
          , Barcelona,
          <source>Spain (Online)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>2296</fpage>
          -
          <lpage>2309</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .coling-main.
          <volume>208</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Macko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Uchendu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lucas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yamashita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pikuliak</surname>
          </string-name>
          , I. Srba,
          <string-name>
            <given-names>T.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Simko</surname>
          </string-name>
          , M. Bielikova,
          <article-title>MULTITuDE: Large-scale multilingual machine-generated text detection benchmark</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>9960</fpage>
          -
          <lpage>9987</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .emnlp-main.
          <volume>616</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          . emnlp-main.
          <volume>616</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Mitchell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khazatsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Finn</surname>
          </string-name>
          , Detectgpt:
          <article-title>Zero-shot machine-generated text detection using probability curvature</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2301</volume>
          .
          <fpage>11305</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schwarzschild</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Cherepanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kazemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Goldblum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geiping</surname>
          </string-name>
          , T. Goldstein,
          <article-title>Spotting llms with binoculars: Zero-shot detection of machine-generated text</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2401</volume>
          .
          <fpage>12070</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Uchendu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ma</surname>
          </string-name>
          , T. Le,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <surname>TURINGBENCH:</surname>
          </string-name>
          <article-title>A benchmark environment for Turing test in the age of neural text generation</article-title>
          , in: M.
          <article-title>-</article-title>
          <string-name>
            <surname>F. Moens</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Specia</surname>
          </string-name>
          , S. W.-t. Yih (Eds.),
          <source>Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2021</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Punta Cana, Dominican Republic,
          <year>2021</year>
          , pp.
          <fpage>2001</fpage>
          -
          <lpage>2016</lpage>
          . doi:
          <volume>10</volume>
          .18653/ v1/
          <year>2021</year>
          .findings-emnlp.
          <volume>172</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mansurov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ivanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tsvigun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. Mohammed</given-names>
            <surname>Afzal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mahmoud</surname>
          </string-name>
          , G. Puccetti,
          <string-name>
            <given-names>T.</given-names>
            <surname>Arnold</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Habash</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , P. Nakov,
          <article-title>M4GT-bench: Evaluation benchmark for black-box machine-generated text detection</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Bangkok, Thailand,
          <year>2024</year>
          , pp.
          <fpage>3964</fpage>
          -
          <lpage>3992</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>218</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Kushnareva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gaintseva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Abulkhanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Magai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tulchinskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barannikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nikolenko</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Piontkovskaya</surname>
          </string-name>
          ,
          <article-title>Boundary detection in mixed AI-human texts</article-title>
          ,
          <source>in: First Conference on Language Modeling</source>
          ,
          <year>2024</year>
          . URL: https://openreview.net/forum?id=kzzwTrt04Z.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tulchinskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kushnareva</surname>
          </string-name>
          , G. Magai,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barannikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nikolenko</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Piontkovskaya</surname>
          </string-name>
          ,
          <article-title>Robust ai-generated text detection by restricted embeddings</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2410</volume>
          .
          <fpage>08113</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>G.</given-names>
            <surname>Gritsai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Voznyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabovoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chekhovich</surname>
          </string-name>
          ,
          <article-title>Are ai detectors good enough? a survey on quality of datasets with machine-generated texts</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2410</volume>
          .
          <fpage>14677</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Crawshaw</surname>
          </string-name>
          <article-title>, Multi-task learning with deep neural networks: A survey</article-title>
          ,
          <year>2020</year>
          . arXiv:
          <year>2009</year>
          .09796.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <article-title>USTC-BUPT at SemEval-2024 task 8: Enhancing machine-generated text detection via domain adversarial neural networks and LLM embeddings</article-title>
          , in: A.
          <string-name>
            <surname>K. Ojha</surname>
            ,
            <given-names>A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Doğruöz</surname>
            ,
            <given-names>H. Tayyar</given-names>
          </string-name>
          <string-name>
            <surname>Madabushi</surname>
            , G. Da San Martino, S. Rosenthal,
            <given-names>A</given-names>
          </string-name>
          . Rosá (Eds.),
          <source>Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Mexico City, Mexico,
          <year>2024</year>
          , pp.
          <fpage>1511</fpage>
          -
          <lpage>1522</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .semeval-
          <volume>1</volume>
          .
          <fpage>217</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>G.</given-names>
            <surname>Gritsai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Voznyuk</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Khabutdinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabovoy</surname>
          </string-name>
          ,
          <article-title>Advacheck at GenAI detection task 1: AI detection powered by domain-aware multi-tasking</article-title>
          , in: F. Alam,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Habash</surname>
          </string-name>
          , I. Gurevych,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Artemova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kutlu</surname>
          </string-name>
          , G. Mikros (Eds.),
          <source>Proceedings of the 1stWorkshop on GenAI Content Detection (GenAIDetect)</source>
          ,
          <source>International Conference on Computational Linguistics</source>
          , Abu Dhabi,
          <string-name>
            <surname>UAE</surname>
          </string-name>
          ,
          <year>2025</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>243</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Greiner-Petter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle, Overview of PAN 2025:
          <article-title>Voight-Kampf Generative AI Detection, Multilingual Text Detoxification, Multi-Author Writing Style Analysis, and Generative Plagiarism Detection</article-title>
          , in: J.
          <string-name>
            <surname>C. de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tsivgun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abassy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mansurov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Ta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Elozeiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. V.</given-names>
            <surname>Tomar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Artemova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Habash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the “VoightKampf” Generative AI Authorship Verification Task at PAN</article-title>
          and
          <article-title>ELOQUENT 2025</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D.</given-names>
            <surname>Macko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Uchendu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lucas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yamashita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pikuliak</surname>
          </string-name>
          , I. Srba,
          <string-name>
            <given-names>T.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Simko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bielikova</surname>
          </string-name>
          , Multitude:
          <article-title>Large-scale multilingual machine-generated text detection benchmark</article-title>
          ,
          <source>in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2023</year>
          , p.
          <fpage>9960</fpage>
          -
          <lpage>9987</lpage>
          . URL: http://dx.doi.org/10.18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>616</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>616</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Xie</surname>
          </string-name>
          , T. Ma, P. Liang,
          <article-title>Composed fine-tuning: Freezing pre-trained denoising autoencoders for improved generalization</article-title>
          , in: M.
          <string-name>
            <surname>Meila</surname>
          </string-name>
          , T. Zhang (Eds.),
          <source>Proceedings of the 38th International Conference on Machine Learning</source>
          , volume
          <volume>139</volume>
          <source>of Proceedings of Machine Learning Research, PMLR</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>11424</fpage>
          -
          <lpage>11435</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolyada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Grahm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elstner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Loebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Continuous Integration for Reproducible Shared Tasks with TIRA.io</article-title>
          ,
          <source>in: Advances in Information Retrieval. 45th European Conference on IR Research (ECIR</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>241</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>D.</given-names>
            <surname>Hendrycks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gimpel</surname>
          </string-name>
          ,
          <article-title>Gaussian error linear units (gelus</article-title>
          ),
          <year>2023</year>
          . URL: https://arxiv.org/abs/1606. 08415. arXiv:
          <volume>1606</volume>
          .
          <fpage>08415</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>