<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">BertT: A Hybrid Neural Network Model for Generative AI Authorship Verification Notebook for PAN at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Zepeng</forename><surname>Wu</surname></persName>
						</author>
						<author role="corresp">
							<persName><forename type="first">Wenyin</forename><surname>Yang</surname></persName>
							<email>cswyyang@fosu.edu.cn</email>
						</author>
						<author>
							<persName><forename type="first">Ma</forename><surname>Li</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Zikai</forename><surname>Zhao</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">Foshan University</orgName>
								<address>
									<settlement>Foshan</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<address>
									<addrLine>September 09-12</addrLine>
									<postCode>2024</postCode>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">BertT: A Hybrid Neural Network Model for Generative AI Authorship Verification Notebook for PAN at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">51A993EC33D5F34F640D792E5680DB5C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:00+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>PAN 2024</term>
					<term>Generative AI Authorship Verification</term>
					<term>BERT</term>
					<term>Transformer1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>With the rapid development and widespread adoption of Large Language Models (LLMs), distinguishing between human-authored and machine-generated texts has become increasingly complex. Although various classification methods have been devised to help identify the origins of texts, they often fail to address the fundamental feasibility and inherent challenges of the task. Building on extensive experience in the field of authorship verification, this study introduces BertT, a novel hybrid model that combines BERT and Transformer technologies, specifically designed for the Generative AI Authorship Verification Task organized in collaboration with PAN and ELOQUENT Labs. This task requires accurately identifying human-authored texts from pairs, one written by a human and the other generated by a machine. Leveraging the deep semantic understanding capabilities of BERT and the efficient sequence processing power of Transformers, our model, BertT, significantly outperforms existing baseline models such as Fast-Detect.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Text classification is a cornerstone of Natural Language Processing (NLP), with authorship verification serving as a pivotal application in this domain. This process is crucial for validating the authenticity of documents, detecting plagiarism, and identifying the origins of articles, thereby preserving the integrity of written content across various fields. The Generative AI Authorship Verification Task at PAN@CLEF 2024 <ref type="bibr" target="#b0">[1]</ref>, which builds upon previous challenges, aims specifically to differentiate between human-authored and machine-generated texts. This task is increasingly pertinent as Large Language Models (LLMs) like GPTs now produce highquality text that closely mimics human writing, thereby presenting substantial challenges in differentiation.</p><p>The utility of authorship verification has been demonstrated in various contexts, underscoring its adaptability and critical importance. For example, Halvani et al. explore the use of compression models for authorship verification, highlighting their effectiveness in digital text forensics without relying on complex machine learning algorithms or extensive feature engineering <ref type="bibr" target="#b1">[2]</ref>. This approach is well-aligned with our need for efficient and scalable solutions to manage the vast amounts of text generated by LLMs. Similarly, Bevendorff et al. have adapted the unmasking method to short texts, significantly reducing the amount of material required for effective authorship verification, thereby making it applicable to more practical scenarios <ref type="bibr" target="#b2">[3]</ref>. Additionally, the challenge of distinguishing between machine-generated and human-authored content is accentuated in the work by Bao et al., who developed Fast-DetectGPT. This model improves the efficiency of detecting machine-generated text through the innovative use of conditional probability curvature, thereby reducing computational costs while maintaining high accuracy <ref type="bibr" target="#b3">[4]</ref>.</p><p>This advancement is particularly relevant to our study as it addresses similar challenges concerning processing efficiency and accuracy. Moreover, the broad application of neural networks in text classification tasks is exemplified by <ref type="bibr">Yang et al. and Yuan et al.</ref> in their respective studies on profiling irony and stereotype spreaders on Twitter. These studies employ RNN and CNN models to classify complex social media content, offering insights into the adaptability of neural network architectures for varied NLP tasks <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b5">[6]</ref>.</p><p>In response to the complexities of authorship verification in the era of LLMs, we developed Bert_T, a novel model that marries the deep semantic understanding capabilities of BERT with the efficient sequential data processing power of Transformer architecture. This model employs a sophisticated contrastive learning approach with an advanced loss function, aimed at enhancing the discrimination between human and machine text. It operates on a dataset formatted in pairs "(text1, text2, label)", training to detect subtle nuances that signify distinct authorship styles. Prior to its effectiveness evaluation, we submitted our model to the TIRA.io platform <ref type="bibr" target="#b6">[7]</ref>, which provides a stringent and controlled testing environment for fair and transparent benchmarking against established baselines. This preliminary submission was crucial for assessing the model's real-world applicability and refining its performance based on unbiased feedback. The Bert_T demonstrated superior performance across several key metrics, achieving a ROC-AUC of 0.967, a Brier score of 0.903, a C@1 of 0.869, an F1 score of 0.869, and an F0.5u of 0.872, culminating in an overall mean score of 0.896. These results significantly surpassed those of other baseline models such as Fast-DetectGPT (Mistral), PPMd, Unmasking, and Fast-DetectGPT, underscoring the Bert_T's enhanced ability to discern between human and machine-generated texts. This success highlights the efficacy of our approach in tackling the complexities of generative AI authorship verification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Dataset</head><p>The dataset for the Generative AI Authorship Verification Task at PAN@CLEF 2024 plays a crucial role in training and validating the efficacy of our Bert_T model. This year, the dataset comprises a diverse array of text genres, reflecting a mix of both real and synthetically generated content. The primary sources of data include news articles, Wikipedia introduction texts, and pieces of fanfiction, which provide a rich variety in style, structure, and complexity. Additionally, PAN participants receive a bootstrap dataset that includes real and fabricated news articles covering various 2021 U.S. news headlines, designed to simulate scenarios that models might encounter in practical applications.</p><p>The data, sourced from contributions by ELOQUENT participants, is meticulously curated to ensure a balanced representation of human and machine-authored texts. The bootstrap dataset is formatted as newline-delimited JSON files, where each file contains a list of articles. These articles are authored either by one or more human authors or entirely by an AI, specifically Google's Gemini Pro model. The dataset structure is pivotal for the task, as it contains pairs of texts where each pair is written on the same topic but by different authors-one human and one machine. The file format for these pairs is demonstrated below: {"id": "gemini-pro/news-2021-01-01-2021-12-31-kabulairportattack/art-081", "text": "..."} {"id": "gemini-pro/news-2021-01-01-2021-12-31-capitolriot/art-050", "text": "..."} Each text pair in the dataset is meticulously labeled with `0` or `1`, indicating whether the texts are from the same author, thereby facilitating supervised learning. The test dataset is provided in a slightly altered format to challenge the model's ability to generalize. Instead of individual files, it is delivered as a single JSONL file where each line contains a pair of texts. The content of this file is arranged such that the identities of the authors are anonymized, and the order of texts scrambled:</p><p>{"id": "iixcWBmKWQqLAwVXxXGBGg", "text1": "...", "text2": "..."} {"id": "y12zUebGVHSN9yiL8oRZ8Q", "text1": "...", "text2": "..."} Participants are tasked with predicting which of the two texts in each pair is human-authored. This setup tests the model's ability to discern subtle linguistic and stylistic nuances that typically distinguish human writing from its AI-generated counterpart. Access to the dataset is regulated via Zenodo, where participants must register and request access using their Tira-registered email, ensuring that the use of this data remains confined to research purposes and that no redistribution occurs. This controlled distribution ensures compliance with copyright regulations and maintains the integrity of the data for academic and developmental uses.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Dataset Preprocessing</head><p>Effective data preprocessing is essential for the robust performance of machine learning models, particularly in tasks involving natural language processing such as authorship verification. For the Generative AI Authorship Verification Task at PAN@CLEF 2024, our preprocessing routine involved several critical steps to enhance the quality and consistency of model inputs.</p><p>Initially, the text normalization process involved removing all punctuation and converting text to lowercase to reduce variability and focus the model's learning on substantive content. This was coupled with the removal of non-alphabetic characters and numerals to ensure that the model trained strictly on textual elements. Following normalization, stopwords-common words that typically do not contribute to the identification of authorship-were removed to minimize data noise and enhance focus on more distinctive text features.After cleaning the texts, the corpus was tokenized into individual words or tokens, which is essential for structuring raw text into a format suitable for machine learning models. The texts were then vectorized using a pre-trained Bert tokenizer, which also standardized the length of tokens through padding and truncation to optimize computational efficiency. To address the challenge of limited training data, we implemented data augmentation techniques to artificially expand the dataset, creating new text pairs from existing ones by subtly modifying texts while preserving their key attributes. This approach helped improve the model's generalization capabilities from training scenarios to realworld applications.</p><p>Throughout the preprocessing stages, we meticulously ensured that the alterations did not compromise the semantic integrity or the stylistic attributes of the texts, which are crucial for authorship identification. This comprehensive preprocessing not only prepared the dataset for effective training of our Bert_T model but also enhanced the model's accuracy in distinguishing between human and machine-generated texts, a critical aspect of the verification task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Network Architecture</head><p>In this study, we introduced Bert_T, a hybrid neural network model that integrates BERT-base for robust feature extraction with a Transformer encoder to handle attention-based dynamics, specifically tailored for distinguishing between human-written and machine-generated texts. We employ the bert-base-uncased model from Hugging Face's Transformers library as our foundational pre-trained BERT layer, leveraging its well-established capabilities in natural language understanding. This layer focuses on the CLS token embedding to capture comprehensive textual context, which is then processed through a Dropout layer to prevent overfitting and enhance generalizability. The Transformer Encoder, equipped with a multi-head attention mechanism, dynamically integrates information across text segments, crucial for identifying subtle linguistic and stylistic nuances. During testing, Bert_T processes each text in a pair independently in a JSONL format, evaluating the likelihood of each text being human-written and comparing these scores to classify texts; the text with the higher score is deemed humanauthored. Optimization of model parameters such as learning rate and batch size, along with the use of Binary Cross-Entropy Loss, fine-tunes the model's accuracy, ensuring it performs effectively on metrics such as ROC-AUC and Brier scores. This configuration enables Bert_T to meet the specific challenges of the Generative AI Authorship Verification Task at PAN@CLEF 2024, demonstrating both innovative theoretical approaches and practical discriminative capabilities, as illustrated in Figure <ref type="figure" target="#fig_0">1</ref>: Bert_T Architecture. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments and Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Experimental Setting</head><p>In our experimental setup for evaluating the Bert_T model's ability to distinguish between human-authored and machine-generated texts, we preprocessed the dataset and divided it into training and testing sets with a 7:3 ratio. The model, a Bert_T, integrates a pretrained BERT base model with a Transformer layer tailored for sequence classification, featuring 768 hidden units, four attention heads, and a linear classifier. Training parameters were meticulously set, with a batch size of 8 and a learning rate of 1e-6 over 300 epochs using the AdamW optimizer on CUDAcapable GPUs to balance computational efficiency and learning depth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Metrics</head><p>Our evaluation framework was meticulously designed to rigorously assess the performance of the Bert_T model across several metrics that reflect its effectiveness in distinguishing between human-authored and machine-generated texts. The model was evaluated using a standard set of metrics that are commonly employed in authorship verification tasks, including ROC-AUC, Brier score, C@1, F1, and F0.5u, along with the arithmetic mean of these metrics to provide a comprehensive overview of performance.</p><p>Performance Metrics: ROC-AUC measures the area under the receiver operating characteristic curve, providing insight into the model's ability to discriminate between classes across all thresholds <ref type="bibr" target="#b7">[8]</ref>. The ROC curve plots the true positive rate (TPR) against the false positive rate (FPR) at various threshold settings. The formula is given by:</p><formula xml:id="formula_0">ROC-AUC = ∫ TPR(𝑡) 1 0 𝑑(FPR(𝑡))<label>(1)</label></formula><p>Brier Score evaluates the mean squared error of the probabilities assigned, indicating the accuracy of probability predictions <ref type="bibr" target="#b8">[9]</ref>. The lower the Brier score, the better, as it reflects a closer proximity to the true outcome. It is calculated as:</p><formula xml:id="formula_1">Brier Score = 1 𝑁 ∑(predicted probability 𝑖 − actual outcome 𝑖 ) 2 𝑁 𝑖=1<label>(2)</label></formula><p>C@1 represents a modified accuracy that treats non-answers (predictions with a confidence score of 0.5) by averaging the accuracy of the remaining cases, thus penalizing uncertainty <ref type="bibr" target="#b9">[10]</ref>. This metric is particularly useful in situations where making no prediction is preferable to making an incorrect prediction. The formula is: </p><formula xml:id="formula_2">𝐶@1 = Number of</formula><p>F1 Score is the harmonic mean of precision and recall, offering a balance between the precision of the classifier and its recall capability <ref type="bibr" target="#b10">[11]</ref>. It is particularly useful in situations where an equal balance between precision and recall is desired. The formula is:</p><formula xml:id="formula_4">𝐹1 = 2 ⋅ Precision × Recall Precision + Recall<label>(4)</label></formula><p>Where Precision = TP TP+FP 𝑎𝑛𝑑 Recall = TP TP+FN . F0.5u is a variant of the F-measure that weights precision more than recall, suitable for scenarios where false positives are more costly than false negatives <ref type="bibr" target="#b11">[12]</ref>. It is calculated using the formula:</p><formula xml:id="formula_5">𝐹0.5𝑢 = (1 + 0.5 2 ) ⋅ Precision × Recall 0.5 2 ⋅ Precision + Recall<label>(5)</label></formula><p>These metrics collectively provided a robust framework for evaluating our model, enabling us to effectively measure its ability to perform authorship verification across different dimensions of accuracy and reliability.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Results</head><p>Our Bert_T model demonstrated robust performance in the PAN 2024 Voight-Kampff Generative AI Authorship Verification task, showcasing substantial effectiveness across several critical metrics. As evidenced in Table <ref type="table" target="#tab_1">1</ref>, Bert_T achieved a ROC-AUC of 0.967, which, while slightly lower than the top-performing Baseline Binoculars at 0.972, reflects a high level of discriminative capability. The Brier score for Bert_T was 0.903, indicating reliable probability predictions of class membership, although it did not surpass the Baseline Binoculars, which scored 0.957. Regarding precision-related metrics, Bert_T recorded scores of 0.869 for both C@1 and F1, and 0.872 for F0.5u, remaining competitive although below the near-perfect scores around the 95th percentile.</p><p>Table <ref type="table" target="#tab_2">2</ref> presents an overview of Bert_T's mean accuracy across nine test set variants, showing considerable stability and less variability in performance compared to other models which displayed more significant fluctuations. Bert_T maintained a minimum accuracy of 0.354 and a maximum of 0.980, with a notable median performance of 0.892, and the 25th and 75th percentiles at 0.864 and 0.896, respectively. These figures underscore Bert_T's robust performance across different testing scenarios, highlighting its efficacy in handling the complex demands of the verification task.</p><p>In terms of competition standings, our submission ranked 20th out of 30 participants on the official PAN 2024 leaderboard. Notably, Bert_T outperformed all but one baseline with a ranking score over all test datasets of 0.608, as detailed on the PAN 2024 leaderboard. This ranking underscores our model ' s competitive edge and its significant discriminative power in a challenging environment filled with diverse and sophisticated entries.</p><p>These results affirm that Bert_T not only embodies theoretical innovation but also exhibits significant practical capabilities in the authorship verification domain. The model's ability to effectively discern between human and machine-generated texts makes it a valuable tool for complex text analysis tasks. Future work will focus on further optimizing model parameters, enhancing feature engineering techniques, and expanding the diversity of the training dataset to boost the model ' s generalizability and performance across varied textual contexts. This continuous improvement aims to refine Bert_T's capabilities for higher detection accuracy and broader application scope in real-world scenarios. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>This paper details the development and evaluation of the Bert_T model, our innovative contribution to the PAN 2024 Voight-Kampff Generative AI Authorship Verification task.</p><p>Combining BERT-based feature extraction with a Transformer encoder for attention processing, Bert_T effectively differentiates between human-written and machine-generated texts. It demonstrated strong performance across various metrics, achieving a ROC-AUC of 0.967 and a Brier score of 0.903, which confirms its reliability in predictions. Despite stiff competition from established baselines, Bert_T maintained consistent performance across different test set variants, with accuracies ranging from a minimum of 0.354 to a maximum of 0.980. This showcases its capability to handle diverse and complex textual scenarios effectively. Moving forward, we plan to further refine Bert_T by optimizing its parameters, enhancing its feature engineering techniques, and expanding its training dataset to cover a broader spectrum of text types and genres. These efforts will not only improve the model's performance in authorship verification tasks but also extend its applicability to a wider range of natural language processing challenges, aiming for higher detection accuracy and broader operational scope.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Bert_T Architecture</figDesc><graphic coords="4,72.00,97.79,451.00,66.90" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>The final performance of our submission on PAN 2024 (Voight-Kampff Generative AI Authorship Verification)</figDesc><table><row><cell>Approach</cell><cell>ROC-AUC</cell><cell>Brier</cell><cell>C@1</cell><cell>F1</cell><cell>F0.5u</cell><cell>Mean</cell></row><row><cell>Bert_T</cell><cell>0.967</cell><cell>0.903</cell><cell>0.869</cell><cell>0.869</cell><cell>0.872</cell><cell>0.896</cell></row><row><cell>Baseline Binoculars</cell><cell>0.972</cell><cell>0.957</cell><cell>0.966</cell><cell>0.964</cell><cell>0.965</cell><cell>0.965</cell></row><row><cell>Baseline Fast-DetectGPT (Mistral)</cell><cell>0.876</cell><cell>0.8</cell><cell>0.886</cell><cell>0.883</cell><cell>0.883</cell><cell>0.866</cell></row><row><cell>Baseline PPMd</cell><cell>0.795</cell><cell>0.798</cell><cell>0.754</cell><cell>0.753</cell><cell>0.749</cell><cell>0.77</cell></row><row><cell>Baseline Unmasking</cell><cell>0.697</cell><cell>0.774</cell><cell>0.691</cell><cell>0.658</cell><cell>0.666</cell><cell>0.697</cell></row><row><cell>Baseline Fast-DetectGPT</cell><cell>0.668</cell><cell>0.776</cell><cell>0.695</cell><cell>0.69</cell><cell>0.691</cell><cell>0.704</cell></row><row><cell>95-th quantile</cell><cell>0.994</cell><cell>0.987</cell><cell>0.989</cell><cell>0.989</cell><cell>0.989</cell><cell>0.990</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>Overview of the mean accuracy over 9 variants of the test set</figDesc><table><row><cell>Approach</cell><cell>Minimum</cell><cell>25-th Quantile</cell><cell>Median</cell><cell>75-th Quantile</cell><cell>Max</cell></row><row><cell>Bert_T</cell><cell>0.354</cell><cell>0.864</cell><cell>0.892</cell><cell>0.896</cell><cell>0.980</cell></row><row><cell>Baseline Binoculars</cell><cell>0.342</cell><cell>0.818</cell><cell>0.844</cell><cell>0.965</cell><cell>0.996</cell></row><row><cell>Baseline Fast-DetectGPT (Mistral)</cell><cell>0.095</cell><cell>0.793</cell><cell>0.842</cell><cell>0.931</cell><cell>0.958</cell></row><row><cell>Baseline PPMd</cell><cell>0.270</cell><cell>0.546</cell><cell>0.750</cell><cell>0.770</cell><cell>0.863</cell></row><row><cell>Baseline Unmasking</cell><cell>0.250</cell><cell>0.662</cell><cell>0.696</cell><cell>0.697</cell><cell>0.762</cell></row><row><cell>Baseline Fast-DetectGPT</cell><cell>0.159</cell><cell>0.579</cell><cell>0.704</cell><cell>0.719</cell><cell>0.982</cell></row><row><cell>95-th quantile</cell><cell>0.863</cell><cell>0.971</cell><cell>0.978</cell><cell>0.990</cell><cell>1.000</cell></row><row><cell>75-th quantile</cell><cell>0.758</cell><cell>0.865</cell><cell>0.933</cell><cell>0.959</cell><cell>0.991</cell></row><row><cell>Median</cell><cell>0.605</cell><cell>0.645</cell><cell>0.875</cell><cell>0.889</cell><cell>0.936</cell></row><row><cell>25-th quantile</cell><cell>0.353</cell><cell>0.496</cell><cell>0.658</cell><cell>0.675</cell><cell>0.711</cell></row><row><cell>Min</cell><cell>0.015</cell><cell>0.038</cell><cell>0.231</cell><cell>0.244</cell><cell>0.252</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Acknowledgements</head><p>This work was supported by grants from the Guangdong-Foshan Joint Fund Project (No. 2022A1515140096) and Open Fund for Key Laboratory of Food Intelligent Manufacturing in Guangdong Province (No. GPKLIFM-KF-202305).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">B</forename><surname>Casals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korenčić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smirnova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ustalov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Zangerle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">On the usefulness of compression models for authorship verification</title>
		<author>
			<persName><forename type="first">O</forename><surname>Halvani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Winter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Graner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th international conference on availability, reliability and security</title>
				<meeting>the 12th international conference on availability, reliability and security</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Generalizing unmasking for short texts</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
	<note>Long and Short Papers</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Fast-detectgpt: Efficient zero-shot detection of machinegenerated text via conditional probability curvature</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">S</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">B</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">Y</forename><surname>Teng</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.05130</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Intelligent Detection Method for Irony and Stereotype Based on Hybird Neural Networks</title>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Yang</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2022 Labs and Workshops, Notebook Papers</title>
				<editor>
			<persName><forename type="first">Guglielmo</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Nicola</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Allan</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Martin</forename><surname>Potthast</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2022-09">September 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Analysis of Irony and Stereotype Spreaders Based On Convolutional Neural Networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2022 Labs and Workshops, Notebook Papers</title>
				<editor>
			<persName><forename type="first">Guglielmo</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Nicola</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Allan</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Martin</forename><surname>Potthast</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2022-09">September 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Continuous Integration for Reproducible Shared Tasks with TIRA</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kolyada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Grahm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elstner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Loebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_20</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-28241-6_20" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 45th European Conference on IR Research (ECIR 2023)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Kruschwitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Caputo</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="236" to="241" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Deep ROC analysis and AUC as balanced average accuracy, for improved classifier selection, audit and explanation</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Carrington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">G</forename><surname>Manuel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">W</forename><surname>Fieguth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="329" to="341" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Modified Brier score for evaluating prediction accuracy for binary outcomes</title>
		<author>
			<persName><forename type="first">W</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Schnellinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistical methods in medical research</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="2287" to="2296" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">A simple measure to assess non-response</title>
		<author>
			<persName><forename type="first">A</forename><surname>Peñas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rodrigo</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">the Journal of machine Learning research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Generalizing unmasking for short texts</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="654" to="659" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
