<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Team karami-sh at PAN: Transformer-based Ensemble Learning for Multi-Author Writing Style Analysis Notebook for the PAN Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Mohammad</forename><forename type="middle">Karami</forename><surname>Sheykhlan</surname></persName>
							<email>mohammadkaramisheykhlan@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Mohaghegh Ardabili</orgName>
								<address>
									<addrLine>Daneshgah St</addrLine>
									<postCode>5619911367</postCode>
									<settlement>Ardabil</settlement>
									<country key="IR">Iran</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Saleh</forename><surname>Kheiri Abdoljabbar</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">University of Tabriz</orgName>
								<address>
									<addrLine>Bahman Boulevard</addrLine>
									<postCode>5166616471</postCode>
									<settlement>Tabriz</settlement>
									<country key="IR">Iran</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mona</forename><forename type="middle">Nouri</forename><surname>Mahmoudabad</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Mohaghegh Ardabili</orgName>
								<address>
									<addrLine>Daneshgah St</addrLine>
									<postCode>5619911367</postCode>
									<settlement>Ardabil</settlement>
									<country key="IR">Iran</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Team karami-sh at PAN: Transformer-based Ensemble Learning for Multi-Author Writing Style Analysis Notebook for the PAN Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">0D08C1CACB5B6DD29BAFED3DF9C42374</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:00+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Ensemble learning</term>
					<term>Multi-author style change detection</term>
					<term>Natural language processing</term>
					<term>Transformers</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Our study addresses the intricate task of detecting style changes within documents authored by multiple individuals. The primary aim is to pinpoint instances where authors transition within the text. This task holds significant importance in author identification, particularly in situations where no comparative texts are available. By discerning variations in writing style, we can unveil instances of plagiarism, identify instances of gift authorship, authenticate claimed authorships, and potentially develop novel technologies to support writing endeavors. Our methodology employs sophisticated ensemble learning techniques, incorporating fine-tuned Bidirectional Encoder Representations from Transformers (BERT), Robustly Optimized BERT Pretraining Approach (RoBERTa) and Efficiently Learning an Encoder that Classifies Token Replacements Accurately (ELECTRA) models for medium and hard subtask and fine-tuned RoBERTa for easy subtask, to effectively address this complex challenge across diverse datasets, such as those found in Reddit comments. Our team's approach achieved F1-scores of 97.2%, 66.4%, and 64.2% for the Easy, Medium, and Hard subtasks, respectively.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Multi-Author Writing Style Analysis is a challenging field aiming to dissect documents authored collaboratively. It involves discerning authorship within a text and pinpointing stylistic shifts, which is essential for plagiarism detection and validating authorship. With the proliferation of digital content, ensuring robust mechanisms to prevent misappropriation becomes increasingly critical. Automated solutions, like Style Change Detection((SCD) algorithms, show promise but require ongoing refinement. Initiatives like PAN@CLEF 2024 <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref> underscore efforts to enhance style analysis techniques.</p><p>In recent years, PAN@CLEF has organized numerous international competitions focusing on SCD tasks, attracting a diverse array of participants who have contributed innovative methodologies. While some participants have relied on conventional natural language processing techniques to address SCD challenges, others have opted for the utilization of deep learning models for feature extraction. Additionally, there exists a cohort of researchers who have leveraged advanced and intricate algorithms, including BERT and its analogous models, alongside collaborative learning techniques.</p><p>Jacobo et al. <ref type="bibr" target="#b2">[3]</ref> employed two feature extraction techniques, Prediction by Partial Matching (PPM) and Term-document matrix (TD), alongside Support Vector Machine (SVM) and Logistic Regression (LR) algorithms for model training. Their experiments across all three datasets revealed that PPM combined with LR achieved the highest F1-score for the easy and medium subtasks, while Term Frequency-Inverse Document Frequency (TF-IDF) coupled with SVM yielded the highest F1-score for the hard subtask. Zi et al. <ref type="bibr" target="#b3">[4]</ref> proposed a method utilizing a combination of convolutional neural networks, BERT, and bidirectional long short-term memory networks for style change detection and author attribution tasks, employing binary classification. Hashemi and Shi <ref type="bibr" target="#b4">[5]</ref> utilized fine-tuning on transformer-based models A RoBERTa and employed data augmentation and ensemble learning techniques. Their models ranked first in two sub-tasks and second in others, showcasing strong performance. Chen et al. <ref type="bibr" target="#b5">[6]</ref> present a study employing comparative learning techniques for analyzing writing style. They enhance sentence segment embeddings produced by a pre-trained model's encoder to ensure closer proximity for similar stylistic sentences and greater separation for dissimilar ones. Utilizing this optimized encoder, they generate sentence embeddings by combining tag data with paragraph sample pairs and classifying them through a full connection layer. Experimental results demonstrate F1-scores of 0.9145, 0.8203, and 0.6755 on Task 1, Task 2, and Task 3 of the official test set, respectively. Kucukkaya et al. <ref type="bibr" target="#b6">[7]</ref> delve into the task of multi-author writing style detection, which involves identifying shifts in writing style within text documents. They frame this challenge as a natural language inference task, pairing consecutive paragraphs. Their strategy emphasizes paragraph transitions and token truncation for input. Using various Transformer-based encoders with warmup training, they submit a model version that surpasses baselines and other proposed versions in experimentation. Specifically, they employ a transition-focused natural language inference approach based on Decoding-enhanced BERT with Disentangled Attention (DeBERTa) with warmup training for easy and medium setups, while opting for the same model without transitions for the hard setup.</p><p>While feature extraction methods like TF-IDF have yielded satisfactory results in author profiling and Authorship Attribution tasks <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>, their performance is not as effective, particularly on the Hard dataset in the SCD task. This is due to the significant similarity observed among sample texts, limiting their effectiveness. Recent progress in this field has shown that BERT and similar language models, known for their complexity with extensive parameters, perform exceptionally well in SCD tasks.</p><p>In this study, we investigated the efficacy of transformer-based models for detecting changes in authorial style at the paragraph level. Initially, we fine-tuned three transformer-based models on the training dataset. Furthermore, we employed Ensemble learning techniques to augment the performance of our approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Task and Datasets</head><p>The datasets provided by PAN@CLEF for writing SCD are categorized into three difficulty levels: Easy, Medium, and Hard. Each level presents unique challenges for detecting shifts in authorship within documents. On the Easy level, documents contain paragraphs covering diverse topics, allowing approaches to utilize topic information effectively in detecting authorship changes. Conversely, the Medium level entails documents with limited topical variety, requiring a stronger emphasis on stylistic analysis to address the detection task. Finally, the Hard level poses the most challenging scenario, where all paragraphs within a document focus on the same topic, demanding approaches to rely solely on stylistic cues for accurate detection.</p><p>Moreover, for each subtask, PAN@CLEF provides distinct datasets comprising multiple documents, each containing paragraphs. Accompanying these datasets are ground truth files, which furnish essential information: the number of authors associated with each document and the identification of consecutive paragraphs where style changes occur, signifying transitions in authorship. These datasets are partitioned into training and validation sets to facilitate experimental setup. In our approach, we treat every pair of consecutive paragraphs as a sample, concatenating them and assigning a label indicating whether a style change occurred between the two paragraphs (labeled as 1) or if they were written by the same author (labeled as 0).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">System Overview</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Data preparation</head><p>For each subtask, our methodology involves concatenating two consecutive paragraphs and assigning them a binary label, thereby framing each subtask as a binary classification problem. In our study, we harness the power of three state-of-the-art transformer-based models: BERT, RoBERTa, and ELECTRA complemented by their corresponding tokenizers. Given the constraint of 512 tokens imposed by the medium-sized models and the rarity of samples exceeding this limit, we adopt a uniform approach by adhering to the 512-token threshold and truncating longer samples as necessary. Additionally, to ensure consistent attention distribution across paragraphs within each sample, we implement a truncation strategy that involves removing tokens from the end of the sequence. This meticulous approach enables us to effectively leverage the capabilities of these advanced models while addressing the practical constraints of our dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Transformer-based Models</head><p>Transformer models have revolutionized natural language processing, capturing complex relationships in text data through self-attention mechanisms. Unlike traditional models, they process all words in a sentence simultaneously, enabling efficient handling of longer sequences. Pre-trained on large corpora, models like BERT and Generative Pre-trained Transformer (GPT) learn rich language representations, adaptable to various tasks with minimal labeled data. Widely used in machine translation, summarization <ref type="bibr" target="#b9">[10]</ref>, and text classification like hate speech detection <ref type="bibr" target="#b10">[11]</ref>, transformers are fundamental in modern NLP systems. In our study, we employed three popular pre-trained transformer models, namely BERT, RoBERTa, and ELECTRA.</p><p>BERT, RoBERTa, and ELECTRA are prominent transformer-based models in NLP. BERT, introduced by Google, utilizes bidirectional training and transformer architecture to capture contextual information effectively. RoBERTa, an improvement upon BERT, optimizes training strategies and scales model size, achieving superior performance on various NLP tasks. ELECTRA, employing a novel pre-training approach, replaces tokens with plausible alternatives and trains a discriminator to distinguish between real and replaced tokens, enhancing efficiency and learning effectiveness.</p><p>Ultimately, we augmented each language model with a binary classification layer to detect changes in writing style. We fine-tuned the models for each subtask using their respective datasets. Additionally, we aggregated the datasets from all three subtasks and fine-tuned the RoBERTa and ELECTRA models on this combined dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Ensemble learning</head><p>Hard Ensemble Learning refers to a sophisticated approach in machine learning where multiple models, known as base learners, are combined to form a stronger, more robust predictive model. Unlike traditional ensemble methods, such as bagging and boosting, which primarily focus on combining diverse but weak models, Hard Ensemble Learning integrates several complex and high-performing models to tackle challenging tasks. This technique leverages the collective intelligence of diverse models, each trained on different aspects of the data, to improve overall predictive accuracy and generalization.</p><p>In this study, we developed five single models for each subtask, all based on three fine-tuned models: BERT, RoBERTa, and ELECTRA. Initially, we fine-tuned all three models with the dataset corresponding to each subtask. Then, we combined the datasets of all three subtasks and fine-tuned the RoBERTa and ELECTRA models on them. Our goal with ensemble learning is to leverage the strengths of each approach to enhance system performance. Detailed specifics will be provided in the next section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Hyperparameter tuning and Evaluation</head><p>In this study, we utilized Google Colaboratory's GPU to fine-tune the BERT, RoBERTa, and ELECTRA models. Due to token constraints, we limited our consideration to 512 tokens. The hyperparameter learning rate for all three transformer models was set to 2e-5. For the dataset associated with the "easy" subtask, we employed an epoch value of 10, while for the other two subtasks, we used 7 epochs. Moreover, for the fine-tuning models chosen for the entire subtasks dataset (RoBERTa, ELECTRA), the epoch value was set to 3. The F1-score is a balanced measure of a model's accuracy, combining precision and recall. Precision measures the accuracy of positive predictions, while recall measures the model's ability to find all positive instances. The macro F1-score averages F1-scores across all classes, providing a balanced evaluation, especially in datasets with class imbalances. Following our experimentation and analysis of results from evaluation sets, we identify the most effective model for each subtask. Subsequently, we execute the selected model on an unseen test set through the TIRA platform <ref type="bibr" target="#b11">[12]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head><p>In this section, we present the results of our experiments on the SCD task. Table <ref type="table" target="#tab_0">1</ref> summarizes the performance of the tested models. We conducted 12 different experiments for each subtask, both with standalone models and ensemble learning. Our findings show that for the "easy" subtask, the pre-trained RoBERTa model achieved the highest macro F1-score on the provided validation dataset. However, for the remaining subtasks, ensemble learning models outperformed others. Specifically, fine-tuning RoBERTa with the entire subtasks dataset (Combined1) and ELECTRA (Combined2), as well as training ELECTRA, RoBERTa, and BERT models on the medium subtask dataset, yielded the highest macro F1-score accuracy. Moreover, for the "Hard" subtask, models trained on RoBERTa and ELECTRA, along with Combined1, delivered the best performance across all experiments. Therefore, for each subtask's test dataset, we selected the approach with the best performance. The results of our work on the unseen dataset are presented in Table <ref type="table">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>This paper provides an overview of our team's performance in the PAN Shared Task for the SCD task. We embarked on a series of diverse experiments to pinpoint the most effective strategy for identifying shifts in writing style within consecutive paragraphs. Initially, we trained five distinct language models, meticulously fine-tuning each to grasp the intricacies of our task. Following this, we delved into an extensive exploration of various combinations of these models, tailoring our approaches to suit the unique demands of each subtask. For the "easy" subtask, we opted to leverage the capabilities of the RoBERTa model exclusively, given its robust performance in preliminary assessments. However, for the more complex subtasks, we adopted an ensemble learning approach, harnessing the collective power of multiple models to tackle the nuanced challenges presented.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Macro F1 Score on the validation dataset. The best result for each dataset is given in bold. Combined1 and Combined2 refer to fine-tuned RoBERTa and ELECTRA on the entire tasks dataset, respectively.</figDesc><table><row><cell>Model</cell><cell>Easy</cell><cell>Medium</cell><cell>Hard</cell></row><row><cell>BERT</cell><cell>90.95</cell><cell>78.3</cell><cell>66.93</cell></row><row><cell>RoBERTa</cell><cell>97.76</cell><cell>82.94</cell><cell>75.84</cell></row><row><cell>ELECTRA</cell><cell>93.86</cell><cell>81.21</cell><cell>75.84</cell></row><row><cell>Combined1</cell><cell>85.45</cell><cell>81.77</cell><cell>64.19</cell></row><row><cell>Combined2</cell><cell>82.66</cell><cell>80.37</cell><cell>65.54</cell></row><row><cell>RoBERTa+BERT+ELECTRA</cell><cell>95.45</cell><cell>82.84</cell><cell>77.08</cell></row><row><cell>RoBERTa+ELECTRA+Combined1</cell><cell>98.18</cell><cell>83.17</cell><cell>77.54</cell></row><row><cell>RoBERTa+ELECTRA+Combined2</cell><cell>95.9</cell><cell>82.89</cell><cell>77.38</cell></row><row><cell>Combined1+Combined2+RoBERTa</cell><cell>90.9</cell><cell>83.12</cell><cell>70.46</cell></row><row><cell>Combined1+Combined2+ELECTRA</cell><cell>89.44</cell><cell>82.28</cell><cell>70.62</cell></row><row><cell>Combined1+Combined2+BERT</cell><cell>89.21</cell><cell>82.68</cell><cell>68.85</cell></row><row><cell cols="2">Combined1+Combined2+RoBERTa+ELECTRA+BERT 95.49</cell><cell>83.57</cell><cell>75.50</cell></row><row><cell>Table 2</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Evaluation of Final Performance on Unseen Test Sets</cell><cell></cell><cell></cell></row><row><cell>Task</cell><cell>Reported F1-score</cell><cell></cell><cell></cell></row><row><cell>Task 1</cell><cell>97.2</cell><cell></cell><cell></cell></row><row><cell>Task 2</cell><cell>66.4</cell><cell></cell><cell></cell></row><row><cell>Task 3</cell><cell>64.2</cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">B</forename><surname>Casals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korenčić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smirnova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ustalov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Zangerle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the Multi-Author Writing Style Analysis Task at PAN 2024</title>
		<author>
			<persName><forename type="first">E</forename><surname>Zangerle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Authorship verification machine learning methods for Style Change Detection in texts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Jacobo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dehesa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rojas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<idno>CEUR-WS.org</idno>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-217.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2652" to="2658" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Style Change Detection Based On Bi-LSTM And Bert</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">Z J</forename><surname>Zia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liua</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-3180/paper-234.pdf" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2022 Labs and Workshops</title>
		<title level="s">Notebook Papers</title>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Enhancing Writing Style Change Detection using Transformer-based Models and Data Augmentation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hashemi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Shi</surname></persName>
		</author>
		<idno>CEUR-WS.org</idno>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-212.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2613" to="2621" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A Writing Style Embedding Based on Contrastive Learning for Multi-Author Writing Style Analysis</title>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Han</surname></persName>
		</author>
		<idno>CEUR-WS.org</idno>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-206.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2562" to="2567" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">ARC-NLP at PAN 23: Transition-Focused Natural Language Inference for Writing Style Detection</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">E</forename><surname>Kucukkaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Sahin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Toraman</surname></persName>
		</author>
		<idno>CEUR-WS.org</idno>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-218.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2659" to="2668" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Cross-domain Authorship Attribution: Author Identification using a Multi-Aspect Ensemble Approach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Rahgouy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Giglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rahgooy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sheykhlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mohammadzadeh</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2380/" />
	</analytic>
	<monogr>
		<title level="m">CLEF 2019 Labs and Workshops</title>
		<title level="s">Notebook Papers</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Losada</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Author profiling: Bot and gender prediction using a multi-aspect ensemble approach</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">B</forename><surname>Giglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rahgouy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rahgooy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">K</forename><surname>Sheykhlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mohammadzadeh</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-2380/paper_231.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2019 -Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Losada</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</editor>
		<meeting><address><addrLine>Lugano, Switzerland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">September 9-12, 2019. 2019</date>
			<biblScope unit="volume">2380</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lapata</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1908.08345</idno>
		<title level="m">Text summarization with pretrained encoders</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Pars-hao: Hate speech and offensive language detection on persian social media using ensemble learning</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">K</forename><surname>Sheykhlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shafi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kosari</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Authorea Preprints</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Continuous Integration for Reproducible Shared Tasks with TIRA</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kolyada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Grahm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elstner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Loebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_20</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-28241-6_20" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 45th European Conference on IR Research (ECIR 2023)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Kruschwitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Caputo</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="236" to="241" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
