<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>WhosAI: A Contrastive Learning Framework for Machine-Generated Text Detection and Attribution</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>(Discussion Paper)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lucio La Cava</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Tagarelli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. Computer Engineering</institution>
          ,
          <addr-line>Modeling, Electronics, and Systems Engineering (DIMES)</addr-line>
          ,
          <institution>University of Calabria</institution>
          ,
          <addr-line>87036 Rende (CS)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>The rapid advancement of Large Language Models (LLMs) has increasingly blurred the line between human and machine-generated text (MGT), raising new societal challenges. As MGT content becomes more widespread and harder to detect, robust identification methods are crucial. In this work, we present WhosAI, a triplet-network contrastive learning framework designed to detect and attribute AIgenerated text. Unlike most existing methods, our framework simultaneously learns semantic similarity representations from multiple text generators, enabling it to efectively handle both detection (human vs. machine) and authorship attribution (identifying the specific generator). Furthermore, WhosAI is model-agnostic and scalable, seamlessly adapting to new LLMs for text-generation by integrating their outputs into its learned embedding space. Experimental results on the TuringBench dataset of 200K news articles demonstrate that our framework excels in both the Turing Test and Authorship Attribution tasks, outperforming all existing methods on the TuringBench leaderboard.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;machine-generated text</kwd>
        <kwd>AI detection</kwd>
        <kwd>AI attribution</kwd>
        <kwd>contrastive learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Recent breakthroughs in artificial intelligence (AI) have significantly advanced natural language
processing (NLP), leading to powerful text generation models capable of producing highly
lfuent and human-like content [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. As text generation models become increasingly realistic,
distinguishing between machine-generated text (MGT) and human writing becomes a pressing
challenge [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Without efective methods for identifying MGT, the preservation of truth,
authenticity, and trustworthiness in online communication is at risk, posing several challenges
to our society. Moreover, the growing presence of MGT in digital channels favors the risk
of misinformation, manipulation, and deception [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Therefore, robust detection mechanisms
are essential to prevent users from unknowingly consuming or spreading false or misleading
information, compromising the integrity of public discourse and decision-making processes.
Furthermore, as MGT closely mimics human language and creativity [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], ethical and societal
concerns arise about authorship, intellectual property rights, and accountability, challenging
established norms in intellectual property law and digital content creation in the absence of
proper mechanisms for identifying the origin of text.
Related work. The aforementioned challenges determined growing interest in detecting
whether and to what extent texts have been generated by humans or machines [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. One
approach dubbed “watermarking” [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ] involves embedding specific signals into generated
texts that remain invisible to humans but are algorithmically detectable. Statistical learning
approaches for detecting the authorship of texts include probabilistic models [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ], log rank
information [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], perplexity [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], discourse motifs [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], and other statistical approaches [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14,
15, 16</xref>
        ]. More recently, deep learning approaches have been proven promising in detecting
or attributing MGT. These include exploiting LLMs to detect generated text [17, 18], using
ChatGPT itself as a detector [19], or combining LLMs with topological aspects [20]. Similarly,
contrastive representation learning has been proven particularly efective in NLP contexts,
such as text classification [ 21, 22], hate-speech detection [23], unveiling intents [24], and MGT
detection [25, 26]. Despite progress in MGT detection research, each of the above mentioned
approaches faces notable challenges. Watermarking methods depend on the possibility to
embed watermarks in text, statistical learning methods often require access to model internals
or information that might be unavailable, and existing contrastive-learning-based methods
typically perform best when trained separately for each generator.
      </p>
      <p>Contributions. In this paper, we discuss WhosAI, a recently proposed contrastive-learning
framework conceived to address binary/multi-class prediction tasks of texts written by humans
or machine text-generation models [27]. The core idea behind WhosAI is to integrate the power
of Transformer-based pretrained language models (PLMs) into a similarity learning framework
optimizing a contrastive triplet loss function to learn deep semantic subspaces that maximize
the cohesiveness among similar texts and the separation between dissimilar texts. Compared to
existing MGT detection methods, WhosAI ofers the following key advantages. First, unlike
watermarking methods, WhosAI does not require editing texts, accessing models’ internals, or
assume specific linguistic features, being able to deal with any type of texts. Second, WhosAI is
conceived to be versatile w.r.t. the particular PLM used at the core of the learning framework,
and is general-purpose, eliminating the need for separate models tailored to specific tasks or
generators.Finally, our contrastive learning framework makes WhosAI model-agnostic and
scalable to the release of new AI text-generators—simply incorporating new data into training
enables generalization to new models. The efectiveness of WhosAI has been demonstrated
through a comprehensive evaluation on the widely recognized TuringBench benchmark dataset,
which includes 200K articles that are either human-written or machine-generated by 19 diferent
AI text-generation models. WhosAI achieves excellent results in terms of both classification
performance and internal validity criteria, outperforming all the methods appearing in the
benchmark’s leaderboard, for both the Turing Test and Authorship Attribution tasks.</p>
      <p>In the remainder of this paper, we summarize and discuss the main findings drawn from the
development and evaluation of WhosAI. For comprehensive details on the design of WhosAI
and in-depth discussion of results, the interested reader is referred to [27].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem Statement</title>
      <p>We are given a set of discrete labels (categories)  = { }=1, with  ≥ 2 , and a collection
of text data objects  = {}=1, such that each text object in  is assigned to one of the
categories in , indicating whether it was authored by a human or a machine. The authorship
of all texts in  are a priori known.</p>
      <p>Our goal is to learn a model, supervisedly trained on ⟨, ⟩, that can predict the category
from  for a given text whose authorship is unknown. Specifically, we tackle the following
problems: () Turing Test (TT), a binary classification task to determine whether the author of
a text is a human or an AI text-generator, and () Authorship Attribution (AA), a multi-class
classification task, that identifies the specific author of a text, choosing between a human or
an AI text-generator. Following literature, our setting does not diferentiate between human
authors in either task, with ‘human’ always corresponding to one class in . Also, the identity
of a particular machine text-generator must be unveiled for the Authorship Attribution task
only, therefore  − 1 categories are available that correspond to either any AI text-generator
(for TT) or a specific AI text-generator (for AA).</p>
    </sec>
    <sec id="sec-3">
      <title>3. The WhosAI Framework</title>
      <p>Overview. WhosAI is a deep learning framework for detecting and attributing open-ended
texts, distinguishing between AI-generated and human-written content.</p>
      <p>Figure 1 illustrates the main components and data flow of the framework. It is trained
on labeled text data, where authorship is categorized as either human or AI-generated, and
consists of three key components: (i) a pretrained language model (PLM) that learns deeply
contextualized text representations in an unsupervised manner, (ii) a Triplet Network that
applies contrastive learning to structure a similarity space for PLM embeddings, and (iii) a
nearest centroid classifier that predicts the authorship of query texts.</p>
      <p>During training, WhosAI learns a deep semantic representation space, where distinct regions
capture the characteristics of human-written and AI-generated texts. Contrastive learning
allows for capturing the underlying data-similarity structure, by grouping embeddings from
the same author while separating those from diferent authors. This learned similarity space
hence facilitates classification by defining clear decision boundaries. In this setting, the nearest
centroid classifier provides an eficient and efective approach to authorship attribution, ensuring
robust predictions for previously unseen texts.</p>
      <p>WhosAI is designed to be versatile and modular. Its versatility lies in the flexibility to select
diferent PLMs as the core of the Triplet Network, experiment with various Triplet Network
architectures, and use alternative instance-based classification models. Additionally, WhosAI
follows a modular design, allowing for targeted enhancements to improve specific aspects of the
framework. These enhancements include: (i) optimizing the contrastive learning component for
better eficiency and generalization, (ii) refining class separation in the learned representation
space to better distinguish diferent text creators, and (iii) increasing robustness by introducing
perturbations to the input textual data [27].</p>
      <sec id="sec-3-1">
        <title>3.1. Training</title>
        <p>Transformer-based Pre-trained Language Models (PLMs) are the well-established NLP
tools to build deeply contextualized text-representation learning models. Given a text data
 ∈ , a token sequence  = [ ,1, . . . ,  ,||] is produced as initial representation of 
through a tokenization process typically associated with a PLM. Each token sequence is deeply
contextualized by mapping it onto a dense, relatively low dimensional space of size  , based on
the PLM. The resulting output is the token embeddings of , denoted as PLM() ∈ R×| |.
Eventually, a pooling function pooling(·) is applied to the token embeddings of each object 
to yield a single embedding vector h of size  :

h = pooling(PLM()) ∈ R .
(1)
Typically, this pooled output is an average embedding over all token embeddings of a data
object. The embeddings h are commonly referred to as sentence embeddings.
Similarity Learning. The deeply contextualized representations produced by a PLM lend
themselves particularly suited to enable semantic comparisons between the input text objects. In
this respect, we leverage the similarity space induced from the sentence embeddings. Similarity
learning aims to train a model to distinguish between similar and dissimilar pairs of objects.
More specifically, if we consider objects whose relative similarity follows a predefined orde—i.e.,
for any triplet of objects, the first object is assumed to be more similar to the second object than
to the third object–the goal becomes to learn a contrastive loss function, so that it favors small
distances between pairs of objects labeled as similar, and large distances for pairs labeled as
dissimilar. This is certainly our case since it is expected that a human-written text to be similar
to another human-written text than an AI-generated text, or texts generated by the same AI
model to be similar to each other than to texts generated from other AI models.</p>
        <p>Contrastive learning is often performed by using a Siamese Network architecture [28], which
contains two PLM instances sharing the same weights while being trained in parallel on two
input objects to compute comparable outputs. When using a contrastive triplet loss, Siamese
Network is commonly referred to as Triplet Network.</p>
        <p>Training process. Our training process starts with mining triplets ⟨(), (), ()⟩ of text
data objects from  to be fed into our triplet network. Such triplets are formed in such a
way that, for a given anchor (), () and () are selected as positive and negative sample,
respectively, i.e., such that () = () and () ̸= (), where symbols (·) are here used to
denote the category associated with an anchor, positive or negative object.</p>
        <p>The embeddings h(), h(), h() of the anchor, positive and negative objects, respectively, are
next computed according to Eq. 1. Note that the text annotations, i.e., associated categories, are
not required when computing the embeddings, since the PLM is an unsupervised learner.</p>
        <p>Given a triplet, the Triplet Network computes the distance between the embedding of the
anchor object and the embedding of the positive object (positive pair), and the distance between
the embedding of the anchor object and the embedding of the negative object (negative pair).
The triplet loss minimizes the distance between an anchor and a positive, both having the same
category, and maximizes the distance between the anchor and a negative of a diferent category:
ℒ =
∑︁
max((h(), h()) − (h (), h()) + , 0)
(2)
where (·, ·) is a distance function and  ∈ R + is a margin between positive and negative pairs.
This loss defines the triplet constraint as the requirement that the distance of negative pairs
should be larger than the distance of positive pairs.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Inference</title>
        <p>At inference time, WhosAI exploits an of-line step that consists in precomputing the centroids
in  for each category  ∈ , defined as c = (1/||) ∑︀∈ h, where  denotes the
subset of  containing data objects of category .</p>
        <p>Given a previously unseen data object , WhosAI computes its embedding h (Eq. 1), which
is then compared to each of the centroids in such a way that  is assigned to the category *
that corresponds to the least distant centroid:
* = arg min=1.. (h, c).
(3)</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Optimizations</title>
        <p>We discuss here a set optimization techniques as enhancements of key components in WhosAI,
namely improved triplet mining, dynamic margin scheduling, and data corruption.
Improving Triplet Mining. A straightforward implementation of the triplet mining process
involves gathering triplets before each training epoch and feeding batches of these triplets into
the Triplet Network, essentially as an “ofline” process. However, this approach might have two
main drawbacks: (i) not all generated triplets may contain the valuable information needed
for minimizing the loss (Eq. 2), and (ii) triplets regarded as “informative” in an earlier stage of
training might quickly become “uninformative” as the model’s weights undergo updates.</p>
        <p>Within this view, it becomes crucial for the triplet mining process to prioritize an online
identification of the most informative triplets for each training epoch. These should be the
most unexpected ones, i.e., triplets that most violate the margin constraints enforced by the
loss function. This strategy can improve the mining process as it enhances the generalization
capabilities and training stability, and it makes training more eficient by avoiding the inclusion
of the uninformative triplets.</p>
        <p>The above requirements can efectively be fulfilled by the pair mining scheme adopted in the
multi-similarity miner method [29]. Essentially, the pair mining consists in sampling informative
pairs through the relative similarity between the negative and positive pairs sharing a common
anchor. More specifically, a negative pair is selected as one having lower distance than the
hardest positive pair (i.e., the one with the highest distance):
(h(), h()) &lt; max (h(), h()) + .</p>
        <p>()</p>
        <p>A positive pair is selected as one having higher distance than the hardest negative pair (i.e.,
the one with the lowest distance):
(4)
(5)
(h(), h()) &gt; min (h(), h()) − .</p>
        <p>()
Dynamic Margin Scheduling. Another improvement we consider is to make the training of
WhosAI progressively harder. Specifically, by dynamically increasing the margin  in our loss
function (Eq. 2), we require the model to focus on harder negative pairs as the training goes on,
in order to produce an enhanced separation between classes.</p>
        <p>To this aim, we revise the loss function with a dynamic margin that follows a linear schedule
dependent on the training step time  ≥ 0 , which is defined as  () =  min + Δ( mod ) , where
 min ∈ R+ is the initial margin,  Δ ∈ R+ denotes the margin increment, and  represents the
step size of the increment. We begin with an initial, relatively low margin,  min, to facilitate
manageable gradients during early optimization; in fact, at early stage, a model can exhibit some
discriminative ability, however, large margins during this stage would lead to excessively large
gradients, hindering learning. As the optimization progresses and the distance constraints are
enforced, the importance of loss-based gradients gradually diminishes. To prevent stagnation,
the margin is periodically increased by  Δ every  training steps. Based on the definition of
 (), WhosAI integrates a dynamic margin scheduling with the triplet loss:
ℒ() =
∑︁
max((h(), h()) − (h (), h()) +  (), 0).
(6)
⟨(),(),()⟩</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Methodology</title>
      <p>Data. We used the publicly available benchmark TuringBench [30, 31], containing 200K news
articles, where 10K are human-written and the others are machine-generated news articles
equally distributed over 19 AI text-generation models. These correspond to diferent sizes and
implementations of GPT-1 [32], GPT-2 [33], GPT-3 [34], GROVER [35], CTRL [36], XLM [37],
XLNET [38], FAIR [39], TRANSFORMER [40], and PPLM [41]. To foster reproducibility, we
followed the pre-defined train, validation and test set splits provided by TuringBench.
Assessment Criteria and Model Settings. To validate the performance of WhosAI in
detecting and attributing AI-generated text, we resort to standard statistics based on the confusion
matrices derived from testing WhosAI predictions w.r.t. the ground-truth under the Turing
Test task and w.r.t. the ground-truth under the Author Attribution task, respectively. These
include the weighted average (i.e., averaging over the support-weighted mean per class) of
precision ( ), recall (), and 1-score (1). We also calculate distance-based quantitative criteria
to measure how well the learned space aligns with the predefined categorization of the training
texts, in terms of compactness within same-category groups of objects and separation between
groups of objects of diferent categories. Following the most widely used approach to sentence
embedding [42], we used BERT [43] as our reference PLM;1 despite being a baseline model,
we will show that this choice is suficient to demonstrate the strong performance achieved by
WhosAI without relying on more complex architectures.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>Here we summarize our main results on the Turing Test (TT) and Authorship Attribution
(AA) tasks, respectively, achieved by WhosAI (best-performing setting) and competitors. For
additional details, the interested reader is referred to [27].</p>
      <p>Turing Test. As shown in Figure 2 (left), according to the TuringBench leaderboard2 there is a
substantial disparity in the 1 scores for TT, which implies that texts from some generators
are more easily detectable than others. By contrast, WhosAI is able to learn a deep semantic
space for the whole set of generators at once achieving an impressive 1 score of 0.999 on the
whole TT test set supplied by the TuringBench benchmark, setting a new best performance
on the TT. Our remarkable 1 score is further corroborated by a qualitative analysis based
on the visualization provided in Figure 2 (right): while at the beginning of the training the
semantic representation directly induced by the PLM does not adequate separate the human
and AI subspaces, the final trained WhosAI shows its ability to learn perfectly to recognize
the two classes for the TT. This couples with the remarkable results (not shown) in terms of
average within-category compactness (0.931) and average across-category separation (-0,808).
Authorship Attribution. Our first finding derives from a comparison between WhosAI results
against those reported on the TuringBench leaderboard for the AA task, whose top-5
bestperforming models are shown in Figure 3 (left). RoBERTa [45] with a multi-class classification
setting turns out to be the best model in the leaderboard for the AA task, with a 1 score of 0.811,
followed by other BERT-based approaches, as well as the oficial OpenAI detector and machine</p>
      <sec id="sec-5-1">
        <title>1https://huggingface.co/google-bert/bert-base-uncased 2Available at https://turingbench.ist.psu.edu/</title>
        <p>Detection method
WhosAI
RoBERTa
BERT
BERTAA
OpenAI detector
SVM (3-grams)
learning-based models. The winner method from the leaderboard is however outperformed by
WhosAI, which achieves a striking average weighted 1 score, precision and recall of 0.990,
thus demonstrating almost perfect capabilities of authorship prediction. As previously found
for the TT task, the striking 1 scores achieved by WhosAI couple with an evidence of highest
cohesiveness (0.938) and separation (-0.012) of the subspaces associated with the various text
authorships, as shown in Fig. 3 (left).</p>
        <p>It is also worth noting that the outstanding performance by WhosAI in the AA task is not
paired by a state-of-the-art sentence-embedding method for semantic-similarity-related tasks
like SBERT [42], based on a Siamese network using BERT at its core: indeed, as shown in Fig. 3
(right), the intra-class cohesiveness and inter-class separation of the semantic space learned by
SBERT are clearly worse than those achieved by WhosAI.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and Future Work</title>
      <p>We tackled the challenge of detecting and attributing AI-generated text through WhosAI, a
novel PLM-based framework that leverages contrastive learning to induce a semantic similarity
space texts written by humans or AI text-generation models. This similarity space is eficiently
exploited at inference time by means of a nearest centroid classifier to predict the authorship of
unlabeled texts. Extensive experimentation on the well-known TuringBench dataset has revealed
state-of-the-art performances of WhosAI on both TT and AA tasks.</p>
      <p>Our ongoing work includes OpenTuringBench [46], an Open-model-based benchmark and
framework for machine-generated text detection and attribution. Moreover, future work
involves extending WhosAI to other text domains, comparing with advanced yet commercially
licensed AI detection tools (e.g., GPTZero), improving training eficiency [ 47], and investigating
explainability aspects of WhosAI.</p>
      <p>Acknowledgements. AT, resp. LLC, was supported by project “Future Artificial Intelligence
Research (FAIR)” spoke 9 (H23C22000860006), resp. project SERICS (PE00000014), both under
the MUR National Recovery and Resilience Plan funded by the EU - NextGenerationEU.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <sec id="sec-7-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>[17] D. Ippolito, D. Duckworth, C. Callison-Burch, D. Eck, Automatic detection of generated
text is easiest when humans are fooled, arXiv:1911.00650 (2019).
[18] V. Verma, E. Fleisig, N. Tomlin, D. Klein, Ghostbuster: Detecting text ghostwritten by large
language models, arXiv:2305.15047 (2023).
[19] A. Bhattacharjee, H. Liu, Fighting Fire with Fire: Can ChatGPT Detect AI-generated Text?,</p>
        <p>SIGKDD Explor. Newsl. 25 (2024) 14–21.
[20] A. Uchendu, T. Le, D. Lee, Toproberta: Topology-aware authorship attribution of deepfake
texts, arXiv preprint arXiv:2309.12934 (2023).
[21] L. Pan, C.-W. Hang, A. Sil, S. Potdar, Improved text classification via contrastive adversarial
training, in: Proc. AAAI Conf., volume 36, 2022, pp. 11130–11138.
[22] J. Chen, R. Zhang, Y. Mao, J. Xu, Contrastnet: A contrastive learning framework for
few-shot text classification, in: Proc. AAAI-IAAI-EAAI Conf., 2022, pp. 10492–10500.
[23] Y. Kim, S. Park, Y.-S. Han, Generalizable implicit hate speech detection using contrastive
learning, in: Proc. COLING Conf., 2022, pp. 6667–6679.
[24] J. Zhang, T. Bui, S. Yoon, X. Chen, Z. Liu, C. Xia, Q. H. Tran, W. Chang, P. Yu, Few-shot
intent detection via contrastive pre-training and fine-tuning, arXiv:2109.06349 (2021).
[25] A. Bhattacharjee, T. Kumarage, R. Morafah, H. Liu, ConDA: Contrastive domain adaptation
for AI-generated text detection, in: Proc. IJCNLP Conf., 2023, pp. 598–610.
[26] A. Bhattacharjee, R. Morafah, J. Garland, H. Liu, Eagle: A domain generalization
framework for ai-generated text detection, arXiv:2403.15690 (2024).
[27] L. L. Cava, D. Costa, A. Tagarelli, Is contrasting all you need? contrastive learning for the
detection and attribution of ai-generated text, in: ECAI 2024 - 27th European Conference
on Artificial Intelligence, 19-24 October 2024, Santiago de Compostela, Spain, volume
392 of Frontiers in Artificial Intelligence and Applications, IOS Press, 2024, pp. 3179–3186.
doi:10.3233/FAIA240862.
[28] J. Bromley, I. Guyon, Y. LeCun, E. Säckinger, R. Shah, Signature verification using a siamese
time delay neural network, in: Proc. NIPS Conf., 1993, pp. 737–744.
[29] X. Wang, X. Han, W. Huang, D. Dong, M. R. Scott, Multi-similarity loss with general pair
weighting for deep metric learning, in: Proc. CVPR Conf., 2019, pp. 5022–5030.
[30] A. Uchendu, Z. Ma, T. Le, R. Zhang, D. Lee, TURINGBENCH: A benchmark environment
for Turing test in the age of neural text generation, in: Findings of the EMNLP Conf., 2021,
pp. 2001–2016.
[31] A. Uchendu, T. Le, K. Shu, D. Lee, Authorship attribution for neural text generation, in:</p>
        <p>Proc. EMNLP Conf., 2020, pp. 8384–8395.
[32] A. Radford, K. Narasimhan, T. Salimans, I. Sutskever, et al., Improving language
understanding by generative pre-training (2018).
[33] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever, et al., Language models are
unsupervised multitask learners, OpenAI blog 1 (2019) 9.
[34] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan,
P. Shyam, G. Sastry, A. Askell, et al., Language models are few-shot learners, Proc. NIPS
Conf. 33 (2020) 1877–1901.
[35] R. Zellers, A. Holtzman, H. Rashkin, Y. Bisk, A. Farhadi, F. Roesner, Y. Choi, Defending
against neural fake news, in: Proc. NIPS Conf., 2019.
[36] N. S. Keskar, B. McCann, L. R. Varshney, C. Xiong, R. Socher, Ctrl: A conditional transformer
language model for controllable generation, arXiv:1909.05858 (2019).
[37] G. Lample, A. Conneau, Cross-lingual language model pretraining, arXiv:1901.07291
(2019).
[38] Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. R. Salakhutdinov, Q. V. Le, Xlnet: Generalized
autoregressive pretraining for language understanding, Proc. NIPS Conf. 32 (2019).
[39] N. Ng, K. Yee, A. Baevski, M. Ott, M. Auli, S. Edunov, Facebook fair’s wmt19 news
translation task submission, arXiv:1907.06616 (2019).
[40] Z. Dai, Z. Yang, Y. Yang, J. Carbonell, Q. Le, R. Salakhutdinov, Transformer-XL: Attentive
language models beyond a fixed-length context, in: Proc. ACL Conf., 2019, pp. 2978–2988.
[41] S. Dathathri, A. Madotto, J. Lan, J. Hung, E. Frank, P. Molino, J. Yosinski, R. Liu, Plug and
play language models: A simple approach to controlled text generation, arXiv:1912.02164
(2019).
[42] N. Reimers, I. Gurevych, Sentence-BERT: Sentence embeddings using Siamese
BERTnetworks, in: Proc. EMNLP-IJCNLP Conf., 2019, pp. 3982–3992.
[43] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional
transformers for language understanding, in: Proc. NAACL-HLT Conf., 2019, pp. 4171–
4186.
[44] L. McInnes, J. Healy, J. Melville, Umap: Uniform manifold approximation and projection
for dimension reduction, 2020. arXiv:1802.03426.
[45] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, V.
Stoyanov, Roberta: A robustly optimized bert pretraining approach, 2019. arXiv:1907.11692.
[46] L. L. Cava, A. Tagarelli, OpenTuringBench: An Open-Model-based Benchmark and
Framework for Machine-Generated Text Detection and Attribution, arXiv:2504.11369
(2025).
[47] F. Scala, S. Flesca, L. Pontieri, Play it straight: An intelligent data pruning technique for
green-ai, in: Discovery Science, Springer Nature Switzerland, Cham, 2025, pp. 69–85.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Jakesch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Hancock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Naaman</surname>
          </string-name>
          ,
          <article-title>Human heuristics for ai-generated language are lfawed</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>120</volume>
          (
          <year>2023</year>
          )
          <article-title>e2208839120</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V. S.</given-names>
            <surname>Sadasivan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Balasubramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feizi</surname>
          </string-name>
          ,
          <article-title>Can ai-generated text be reliably detected?</article-title>
          ,
          <source>arXiv:2303.11156</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Can</surname>
            <given-names>LLM</given-names>
          </string-name>
          -Generated Misinformation Be Detected?,
          <source>arXiv:2309.13788</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L. L.</given-names>
            <surname>Cava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tagarelli</surname>
          </string-name>
          , Open Models, Closed Minds?
          <article-title>On Agents Capabilities in Mimicking Human Personalities through Open Large Language Models</article-title>
          ,
          <source>in: Proc. AAAI</source>
          <year>2025</year>
          , AAAI Press,
          <year>2025</year>
          , pp.
          <fpage>1355</fpage>
          -
          <lpage>1363</lpage>
          . URL: https://doi.org/10.1609/aaai.v39i2.
          <fpage>32125</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jawahar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdul-Mageed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. V.</given-names>
            <surname>Lakshmanan</surname>
          </string-name>
          ,
          <article-title>Automatic detection of machine generated text: A critical survey</article-title>
          , arXiv:
          <year>2011</year>
          .
          <volume>01314</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-N.</given-names>
            <surname>Chuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <article-title>The science of detecting llm-generated text</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>67</volume>
          (
          <year>2024</year>
          )
          <fpage>50</fpage>
          -
          <lpage>59</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kirchenbauer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geiping</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Katz</surname>
          </string-name>
          , I. Miers,
          <string-name>
            <given-names>T.</given-names>
            <surname>Goldstein</surname>
          </string-name>
          ,
          <article-title>A watermark for large language models</article-title>
          ,
          <source>in: Proc. ICML Conf</source>
          .,
          <year>2023</year>
          , pp.
          <fpage>17061</fpage>
          -
          <lpage>17084</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Yoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kwak</surname>
          </string-name>
          ,
          <article-title>Robust multi-bit natural language watermarking through invariant features</article-title>
          ,
          <source>in: Proc. ACL Conf</source>
          .,
          <year>2023</year>
          , pp.
          <fpage>2092</fpage>
          -
          <lpage>2115</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>E.</given-names>
            <surname>Mitchell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khazatsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Finn</surname>
          </string-name>
          , Detectgpt:
          <article-title>Zero-shot machinegenerated text detection using probability curvature</article-title>
          ,
          <source>in: Proc. ICML Conf., PMLR</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>24950</fpage>
          -
          <lpage>24962</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Qiu</surname>
          </string-name>
          , Seqxgpt:
          <article-title>Sentence-level ai-generated text detection</article-title>
          ,
          <source>arXiv:2310.08903</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Y.</given-names>
            <surname>Zhuo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          , Detectllm:
          <article-title>Leveraging log rank information for zero-shot detection of machine-generated text</article-title>
          ,
          <source>arXiv:2306.05540</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Vasilatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rahwan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maniatakos</surname>
          </string-name>
          ,
          <article-title>Howkgpt: Investigating the detection of chatgpt-generated university student homework through context-aware perplexity analysis</article-title>
          ,
          <source>arXiv:2305.18226</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z. M.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Raheja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <article-title>Threads of subtlety: Detecting machinegenerated texts through discourse motifs</article-title>
          ,
          <source>arXiv:2402.10586</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Strobelt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rush</surname>
          </string-name>
          , GLTR:
          <article-title>Statistical detection and visualization of generated text</article-title>
          ,
          <source>in: Proc. ACL Conf.: System Demonstrations</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>111</fpage>
          -
          <lpage>116</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>E.</given-names>
            <surname>Tulchinskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Laida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cherniavskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nikolenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Burnaev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barannikov</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Piontkovskaya</surname>
          </string-name>
          ,
          <article-title>Intrinsic dimension estimation for robust detection of AI-generated texts</article-title>
          ,
          <source>in: Proc. NIPS Conf</source>
          .,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Venkatraman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Uchendu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          , Gpt-who:
          <article-title>An information density-based machinegenerated text detector</article-title>
          ,
          <source>arXiv preprint arXiv:2310.06202</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>