<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Enhancing Authorship Verification using Sentence-Transformers</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Momen Ibrahim</string-name>
          <email>es-momen.ibrahim2019@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ahmed Akram</string-name>
          <email>eng-ahmed.akram2018@alexu.edu.eg</email>
          <email>es-mostafahmed22@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammed Radwan</string-name>
          <email>es-mohamed.radwan2000@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rana Ayman</string-name>
          <email>es-rana.hussein2023@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mustafa Abd-El-Hameed</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nagwa El-Makky</string-name>
          <email>nagwamakky@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marwan Torki</string-name>
          <email>mtorki@alexu.edu.eg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer and Systems Engineering Department, Faculty of Engineering, Alexandria University</institution>
          ,
          <addr-line>Alexandria</addr-line>
          ,
          <country country="EG">Egypt</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <abstract>
        <p>Authorship verification is a growing field of research that aims to determine whether two texts were written by the same author or diferent authors. In this paper, we describe our system for the PAN@CLEF 2023 Authorship Verification challenge [ 1] which requires solving the task on a cross-Discourse Types and open-set collection of essays (written discourse), emails (written discourse), Interviews (spoken discourse), and Speech transcriptions (spoken discourse). We use Sentence-Transformers which is a popular framework for generating sentence embedding based on the idea of using pre-trained language models, such as BERT (Bidirectional Encoder Representations from Transformers) to capture the semantic features of diferent authors' documents. As a result of studying multiple sentence-transformers, we selected all-MiniLM-L12-v2 model and achieved the first rank in PAN 23 with an overall score of 62.3% on the testing set.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Authorship Verification</kwd>
        <kwd>pre-trained language models</kwd>
        <kwd>Sentence-Transformers</kwd>
        <kwd>SBERT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Authorship Verification (AV) is a specialized area in the field of digital texts that focuses on
analyzing and comparing the unique stylistic and linguistic characteristics present in multiple
texts. The primary objective of AV is to ascertain whether these texts were authored by the
same individual. Its significance has been magnified in recent times due to the abundance of
digital content accessible through digital libraries, online journalism platforms, and social media
networks.</p>
      <p>AV serves a crucial purpose in various domains where automatic verification of document
authorship is essential. With the exponential growth of digital texts available online, the need
to accurately determine the authorship of documents has become increasingly prominent. This
is particularly relevant in contexts where the evaluation of researchers relies on the impact
and quantity of their publications. Similarly, public figures often face scrutiny based on their
social media posts. In such scenarios, AV ofers a valuable tool to establish the authenticity and
authorship of documents.</p>
      <p>The digital landscape, encompassing digital libraries, online journalism, and social media
platforms, has witnessed a surge in textual data. However, along with this surge, there has also
been an upswing in online crimes. AV plays a pivotal role in addressing these challenges by
identifying instances of fraudulent activities, such as phishing emails, and detecting cases of
plagiarism. By analyzing stylistic features embedded in texts, including writing style, genre,
temperament, sentiment, native language, and gender, AV provides insights into the unique
attributes and traits of authors.</p>
      <p>In summary, AV is a specialized field that utilizes advanced techniques to analyze and compare
the distinctive patterns and features present in multiple texts, with the aim of determining
whether they share a common author. The increasing availability of digital texts and the need
for accurate document authorship verification in various domains underscore the significance
of AV in today’s digital landscape.</p>
      <p>
        In the authorship verification task at PAN@CLEF-2023 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] they introduced more challenging
scenarios where each author verification case considers texts from diferent discourse types.
This edition focuses on cross-discourse type authorship verification, including both written
(essays and emails) and spoken language (interviews and speech transcriptions). The corpus
consists of texts from around 100 individuals, all native English speakers aged 18-22. The topics
of the texts are unrestricted, and the level of formality may vary within each discourse type.
      </p>
      <p>
        Most of the documents in the training set of PAN23 are emails, interviews and
speechtranscriptions which are relatively short documents. This inspired us to use Sentence-Transformers
specially SBERT [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] due to their power in semantic similarity tasks in addition to their eficiency.
We evaluated diferent pre-trained models for their quality to embedded sentences.
      </p>
      <p>As, the used models are trained for short sentences, this directed us to chunk the documents
in the pairs to fit in the max-sequence length of the used pre-trained models. We tried diferent
chuncking techniques. These chunking techniques will be explained in section 4.2.</p>
      <p>This paper is structured as follows: Section 2 provides a background and related work on
authorship verification. Section 3 describes the dataset used in the PAN@CLEF 2023 challenge,
including the diferent discourse types and the number of texts in each category. Section 4
explains the preprocesing step on the dataset and the chunking technique for handling long
texts. Section 5.1 presents the methodology used in our system, including the use of
SentenceTransformers for generating sentence embeddings. Section 6 discusses experiments on other
models, including a model that uses only stylometric features. Section 7 discusses the results
and performance of these models on the validation set. Finally, Section 8 concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background and Related work</title>
      <p>
        Reimers and Gurevych (2019) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] propose a new method for generating sentence embeddings,
which are vector representations of sentences that capture their semantic meaning. The authors
use a Siamese neural network architecture, based on the BERT model, to generate embeddings
that are optimized for use in tasks such as semantic similarity and paraphrase detection. The
model is trained on a large corpus of text data, and the resulting embeddings are shown to
outperform previous state-of-the-art methods on several benchmark datasets. The authors
also introduce a new technique for fine-tuning the model on smaller datasets, which further
improves its performance on specific tasks. In PAN@CLEF 2021, Peng et al. (2021) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] propose
a method that utilizes a pre-trained model to encode text information to solve the authorship
verification in the PAN@CLEF 2021. To resolve the problem of long text encoding, the method
proposed is to split long texts into short texts that a pre-trained model, BERT, can encode.
The classification model achieved the highest c@1 and F1-score on the small dataset of PAN
Authorship Verification datasets. Accordingly, the approach described can encode long text
information eficiently in long text pairs.
      </p>
      <p>
        Stylometric feature extraction is another approach proposed by, Weerasinghe et al. (2021) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
The authors demonstrate the impact of diferent feature representations on the performance of
the proposed method of computing the absolute diference between the feature vectors as input
to the logistic regression classifier for each document pair, they evaluate several feature sets,
including word and character n-grams, punctuation usage, and syntactic features.
      </p>
      <p>
        BERT-like transformers for authorship verification also perform well in this task. Manolache
et al., (2021) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] explore the use of them for authorship verification, The authors evaluate several
transformers on the PAN-2020 dataset and find that they achieve high performance, but also
they show through Integrated Gradients XAI technique that, they rely on topical clues rather
than stylistic features. To address this issue, the authors propose new splits for the dataset that
ensure topic and author diversity and show that they improve the models’ domain generalization
ability. The authors also introduce a new dataset, DarkReddit, that contains texts from diferent
domains and genres, and use it to test the models in a low-data regime.
      </p>
      <p>
        Galicia et al.,(2022)[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] proposed a Graph-based Siamese Network approach for the authorship
verification task at PAN 2022 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. They also introduced a way to split PAN22 dataset to ensure
that the training and validation sets are both balanced and author-disjoint. This is important to
avoid misleading results that may arise from data leakage or overfitting.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Data Description</title>
      <p>The evolution of authorship verification tasks in previous editions of PAN@CLEF competitions
is as follows. Initially, the task focused on comparing writing styles in diferent languages and
genres. Later, cross-domain authorship verification using fan-fiction texts was explored and
found to be feasible across diferent fandom. In the 2022 edition, more challenging scenarios
were introduced, involving cross-DT authorship verification, where texts belonged to diferent
discourse types (DTs), all within the realm of written language. The current edition focuses on
cross-discourse type authorship verification, where both written language (essays and emails)
and spoken language (interviews and speech transcriptions) are included as discourse types.</p>
      <p>The dataset provided by PAN is a collection of English texts from diferent Discourse Types
(DT). The goal of the task is to determine if two texts from diferent DTs are written by the same
author. The dataset has a tag for each pair of texts indicating the authorship, the author, and
the discourse type of each text. Since we augment the training set of PAN 23 by the training set
of PAN 22. The next subsections describe both of them.</p>
      <sec id="sec-3-1">
        <title>3.1. Authorship Verification Dataset at PAN 2022</title>
        <p>In that dataset the organizers provide cross-DT authorship verification cases using DTs
corresponding to written language.</p>
        <p>The discourse types and their corresponding number of texts in the train dataset are:
• Essays (written discourse): 2986
• Emails (written discourse): 10116
• Text messages (written discourse): 9446
• Business memos (written discourse): 1980
The corpus consists of texts from approximately 100 individuals who are native English speakers
and share a similar age range of 18-22. The text samples cover a wide range of topics, without
any restrictions, and within each discourse type, there can be variations in the level of formality.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Authorship Verification Dataset at PAN 2023</title>
        <p>In that dataset the organizers provide cross-DT authorship verification cases using DTs not
only corresponding to written language but also to spoken language.</p>
        <p>The discourse types and their corresponding number of texts in the train dataset are:
• Essays (written discourse): 2594
• Emails (written discourse): 7054
• Interviews (spoken discourse): 6090
• Speech transcriptions (spoken discourse): 1934
And also, the corpus consists of texts from approximately 100 individuals who are native English
speakers and share a similar age range of 18-22. The text samples cover a wide range of topics,
without any restrictions, and within each discourse type, there can be variations in the level of
formality.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Data Augmentation</title>
        <p>Based on the strong similarity between the two tasks of PAN 22 and PAN 23, and because the
number of pairs in the new task of PAN 23 is too small, we decided to combine both train
datasets into one train dataset.
• Essays (written discourse): 5580
• Emails (written discourse): 17170
• Text messages (written discourse): 9446
• Business memos (written discourse): 1980
• Interviews (spoken discourse): 6090
• Speech transcriptions (spoken discourse): 1934
And the new number of combinations in that combined dataset are shown in Table 4.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Data Preprocessing</title>
      <sec id="sec-4-1">
        <title>4.1. Data Splitting</title>
        <p>
          This section describes the pre-processing operations we did on the combined dataset.
We used an author-disjoint method similar to the one introduced in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] to split the dataset into
training and validation sets for our model evaluation. This means that no text from one set
has the same author as any text from another set. Since the dataset had 21,100 problems and
only 112 authors, this method resulted in unbalanced sets, with more positive problems than
negative ones. To address this issue, we generated new negative instances by pairing texts
from diferent authors. We followed these steps: Let A and B be the subsets of texts from each
set grouped by author. We obtained positive and negative pairs by computing the Cartesian
products P = A × A and N = A × B respectively. Then, we removed pairs that had the same
DT, and finally randomly sampled positive and negative pairs from P and N sets to balance the
training and validation sets. The final dataset had an equal number of true and false problems.
The total number of problems and the number of problems in the positive class, the number of
texts and authors on each partition are shown in Table 5.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Texts Chunking</title>
        <p>The sentence transformer is based on a Siamese Network that takes two texts (documents)
as input. The length of document 1 and/or document 2 can be larger than the
max-sequencelength,K, allowed by the sentence -transformer model. So, for training, we chunk each document
into a set of chunks , each of size K. We combine the first 2 chunks of the 2 documents, the
second 2 chunks, etc. This creates for each original pair, M pairs, where M is the number of
chucks of the larger document. The chunking at evaluation time is described in section 5.2.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Model Configuration</title>
      <sec id="sec-5-1">
        <title>5.1. Model Architecture</title>
        <p>We chose all-MiniLM-L12-v2 as our model. It is a Siamese network, based on BERT-like
pretrained model. It belongs to the family of sentence transformers. It uses a contrastive learning
objective and has a max sequence length of 256 tokens. It is a lightweight model with a size
of 120 MB. However, it provides high-quality semantic similarity in addition to its eficiency.
Figure 1 shows the model architecture at training and at inference times.</p>
        <sec id="sec-5-1-1">
          <title>5.1.1. Embedding Layer</title>
          <p>It is the first layer of the model, responsible for converting input tokens into dense vector
representations (embeddings). BERT-like models typically use a combination of token
embeddings, segment embeddings, and position embeddings. The token embeddings are learned
representations specific to each token in the model’s vocabulary. Segment embeddings are used
to distinguish diferent sentences, and position embeddings give the model a sense of the order
of words in a sentence since transformers don’t have a built-in understanding of sequence order.</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>5.1.2. Transformer Encoder Layers</title>
          <p>These are the core of the model. Each transformer encoder layer consists of two sub-layers:
• Self-Attention Mechanism (Multi-Head Attention): The purpose of the self-attention
mechanism is to compute a representation of each word that takes into account the
influence of other words in the sentence.</p>
          <p>In each attention head, for every word, the model calculates a score (attention score) that
signifies how much focus should be placed on other words. These scores determine how
much each word will contribute to the final representation of the current word.
The mechanism performs the following steps:
– Query, Key, and Value Vectors: Each input word is transformed into Query, Key,
and Value vectors using learned linear transformations.
– Attention Score: An attention score is computed for each pair of words. This score
is calculated as the dot product of the Query vector of the current word and the Key
vector of the other word, followed by a scaling operation (dividing by the square
root of the dimension of the key vector).
– Softmax Normalization: The attention scores are then passed through a softmax
function to obtain the weights, which ensures they are positive and sum to 1.
– Weighted Sum: Finally, a weighted sum of the Value vectors of all words is
computed, where the weights are the softmax-normalized attention scores.</p>
          <p>This process is done in parallel in each attention head. Each head may potentially learn
to pay attention to diferent types of connections between words.</p>
          <p>The output from each head is concatenated and linearly transformed to produce the final
output of the multi-head attention layer.
• Feed-Forward Neural Network: The output from the self-attention mechanism for each
position passes through this layer. It’s a simple feed-forward neural network that consists
of two fully connected layers with a ReLU activation function in between. This network
doesn’t have recurrent or convolutional connections, and it operates independently on
each position, treating each position identically. This is where most of the model’s
parameters reside.</p>
          <p>Apart from these, there are also residual connections around both the self-attention and FFNN
components, followed by layer normalization. The residual connections help in preventing the
vanishing gradient problem and enable the model to be deeper. Layer normalization is used to
stabilize the network’s learning process and reduce training time.</p>
          <p>These components work together to create a powerful model that can capture complex
patterns in the input data. The stacking of multiple layers allows the model to learn more
abstract and high-level features as the information propagates up the layers.</p>
        </sec>
        <sec id="sec-5-1-3">
          <title>5.1.3. Pooler Layer</title>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Evaluation</title>
        <p>Once the transformer layers have processed the input, the final step in the model architecture
is to convert the output into a fixed-size sentence embedding. This is achieved by applying a
pooling operation to the output of the transformer layers. The output is a single vector that
represents the entire input sentence.</p>
        <p>As can be seen from the model architecture (Figure 1) the output at evaluation time is the
cosine similarity of two chunks from the two documents. At the evaluation, the first chunk of K
tokens of each document is compared, the second grouping of K tokens from each document
is compared, etc. Finally, we take the average of the M scores as the final cosine similarity.
Since the cosine similarity range is from -1 to 1, we re-scale the score to be from 0 to 1 using
the approach proposed in the baseline model of PAN 22. During training, we optimize two
threshold values p1 and p2. All evaluation scores lower than p1 correspond to negative answers,
and all evaluation scores greater than p2 are scaled to positive answers. The remaining scores
(between p1 and p2) are set to 0.5 and are left unanswered.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Hyper-parameter Tuning</title>
        <p>Hyperparameter tuning plays a crucial role in optimizing the performance of the model. In this
section, we outline the hyperparameters used in our experiments and describe the process of
tuning them.</p>
        <sec id="sec-5-3-1">
          <title>5.3.1. Model Configuration</title>
          <p>We employed the "all-MiniLM-L12-v2" model architecture (with 256 tokens as max-seq-length)
for our experiments. This model has shown promising results in various natural language
processing tasks and is pre-trained on a large corpus of text data.</p>
        </sec>
        <sec id="sec-5-3-2">
          <title>5.3.2. Hyper-parameters</title>
          <p>The list below includes the hyper-parameters considered, together with the values selected after
ifne tuning, using the validation set.</p>
          <p>• Chunk Length: We set the chunk length to 256, which determines the maximum length
of text chunks used during training and evaluation.
• Batch Size: The batch size was set to 32, which determines the number of training
samples processed in parallel during each iteration.
• Distance Metric: We used the cosine distance metric to measure the similarity
between text embeddings.
• Loss Function: We utilized the ContrastiveLoss function as the loss function for
training the model.
• Epochs: We trained the model for 189 epochs, representing the number of times the
entire dataset was iterated during training.
• Scheduler: The learning rate scheduler was set to ’warmuplinear’, which gradually
increases the learning rate during the warm-up phase and then linearly decays it.
• Warmup Steps: The warm-up steps were set to 0, indicating that no warm-up phase
was employed.
• Learning Rate: We set the initial learning rate to 2e-05, which determines the step
size during gradient descent optimization.
• eps: The epsilon value was set to 1e-06, which prevents division by zero in certain
calculations.
• Thresholds P1&amp;P2: The thresholds values that were used to leave dificult cases
unanswered as explained in section 5.2 as, they were set to 0.45 and 0.54 respectively.</p>
        </sec>
        <sec id="sec-5-3-3">
          <title>5.3.3. Hyper-parameter Tuning Process</title>
          <p>The process of hyper-parameter tuning involved a combination of manual exploration and
iterative experimentation. We started with initial values based on previous studies and gradually
refined them through empirical evaluation.</p>
          <p>We performed a grid search on the validation set to explore diferent combinations of
hyperparameters. For each combination, we trained the model and evaluated its performance using
the evaluation metrics in section 5.2, such as area under the curve (AUC), C@1 score, F1 score,
Brier score, and overall performance.</p>
          <p>Based on the evaluation results, we iteratively adjusted the hyper-parameters to improve the
model’s performance. This process involved experimenting with diferent values for specific
hyper-parameters while keeping others constant. We repeated this iterative process until we
achieved satisfactory performance on the validation set.</p>
        </sec>
        <sec id="sec-5-3-4">
          <title>5.3.4. Performance Comparison</title>
          <p>To assess the impact of hyper-parameter tuning, we compared the performance of the model
before and after tuning. The evaluation metrics mentioned in section 5.2 were used to measure
and compare the performance across diferent hyper-parameter configurations.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Experiments on Other Models</title>
      <p>
        We performed experiments on two other models. The first is nli-distilroberta-base-v2 , which
is a sentence-transformer model based on distilroberta. It has a max-sequence-length of 512
tokens. The second model is the Stylometric model proposed in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The results of these models
on PAN 23 validation set are given in section 7.
      </p>
    </sec>
    <sec id="sec-7">
      <title>7. Results</title>
      <p>
        So, the best one is the all-MiniLM-L12-v2 model, which was selected for the submissions.
The performance of our three submitted runs on the PAN23 dataset is presented in Table 7,
according to the TIRA [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] website.
0.562
0.562
0.55
      </p>
      <p>F1-score</p>
      <p>We submitted three runs on TIRA website: ‘resolving-globe’,‘reduced-graph’, and
‘goldenottoman’. All of them use same version of the trained model (all-MiniLM-L12-v2). The first
two runs are the same but, the second run has additional option to take p1 and p2 [described
in section 5.2] as input so that we can modify them. And both of them have the same default
values and they use the chunking technique described in section 4.2. The third run was to see
the efect of soft-chunking on the input data so that there is no sentence split between two
chunks unless it has a number of tokens larger than the max sequence length of the model [256
in our case].</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>In this work, we presented the efectiveness of sentence transformers in the authorship
verification problem. Based on the result we found that the all-MiniLM-L12-v2 model achieved the best
performance on the validation dataset compared to other models like nli-distilroberta-base-v2
and Stylometric model. These results highlight the efectiveness of sentence transformers in
tackling authorship verification, showcasing the potential of the all-MiniLM-L12-v2 model
for improving the accuracy of such tasks.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kredens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pezik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Heini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Overview of the Authorship Verification Task at PAN 2023, in: CLEF 2023 Labs and Workshops, Notebook Papers, CEUR-WS</article-title>
          .org,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Borrego-Obrador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chinea-Ríos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Franco-Salvador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Heini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kredens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pęzik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wolska</surname>
          </string-name>
          , , E. Zangerle, Overview of PAN 2023:
          <article-title>Authorship Verification, Multi-Author Writing Style Analysis, Profiling Cryptocurrency Influencers, and Trigger Detection</article-title>
          , in: A.
          <string-name>
            <surname>Arampatzis</surname>
            , E. Kanoulas,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Tsikrika</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          <string-name>
            <surname>Stefanos Vrochidis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Aliannejadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Vlachos</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          , arXiv preprint arXiv:
          <year>1908</year>
          .
          <volume>10084</volume>
          (
          <year>2019</year>
          ). URL: https://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>X</surname>
          </string-name>
          . Sun,
          <article-title>Encoding text information by pre-trained model for authorship verification</article-title>
          .,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>2103</fpage>
          -
          <lpage>2107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Weerasinghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Greenstadt</surname>
          </string-name>
          ,
          <article-title>Feature vector diference based authorship verification for open-world settings</article-title>
          .,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>2201</fpage>
          -
          <lpage>2207</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Manolache</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Brad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Burceanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barbalau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Popescu</surname>
          </string-name>
          ,
          <article-title>Transferring bertlike transformers' knowledge for authorship verification</article-title>
          ,
          <source>arXiv preprint arXiv:2112.05125</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Martinez-Galicia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Embarcadero-Ruiz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ríos-Orduña</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gómez-Adorno</surname>
          </string-name>
          ,
          <article-title>Graphbased siamese network for authorship verification (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chinea-Ríos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Franco-Salvador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Heini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Körner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kredens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pęzik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          , et al.,
          <source>Overview of pan</source>
          <year>2023</year>
          :
          <article-title>Authorship verification, multi-author writing style analysis, profiling cryptocurrency influencers, and trigger detection</article-title>
          ,
          <source>in: Advances in Information Retrieval: 45th European Conference on Information Retrieval</source>
          ,
          <string-name>
            <surname>ECIR</surname>
          </string-name>
          <year>2023</year>
          , Dublin, Ireland, April 2-
          <issue>6</issue>
          ,
          <year>2023</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>III</given-names>
          </string-name>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>518</fpage>
          -
          <lpage>526</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolyada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Grahm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elstner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Loebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Continuous Integration for Reproducible Shared Tasks with TIRA.io</article-title>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval. 45th European Conference on IR Research (ECIR</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>241</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>