<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Team Liuc0757 at PAN: A Writing Style Embedding Method Based on Contrastive Learning for Multi-Author Writing Style Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chang Liu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhongyuan Han</string-name>
          <email>hanzhongyuan@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Haoyang Chen</string-name>
          <email>hoyo.chen.i@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Qingbiao Hu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Foshan University</institution>
          ,
          <addr-line>Foshan</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>This paper explores the identification of authorial style shifts in multi-author documents to facilitate further author identification. The model is proposed to employ contrastive learning techniques to analyze writing styles and optimized the sentence segment embedding output from the encoder of a pre-trained model. This optimization enables the encoder to produce more similar vector representations in space for sentences with similar styles while widening the distance between the embedding representations of paragraphs with diferent styles. By utilizing the contrastive learning-based encoder to generate sentence embeddings through an analysis of labeled data combined with paragraph sample pairs, we classified them using a fully connected layer. Experimental results demonstrate that we achieved F1 scores of 0.696, 0.717, and 0.503 on Task 1, Task 2, and Task 3 of the oficial test set, respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Style Change Detection</kwd>
        <kwd>Contrastive Learning</kwd>
        <kwd>Sentence Embedding</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the realm of text analytics, discerning writing style shifts in multi-author documents is a complex yet
fascinating task. This analysis not only verifies textual integrity but also identifies potential plagiarism.
Recently, following the advancement of PAN series tasks with limited topic diversity, style analysis has
garnered greater attention. In this context (PAN 2024 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]), we shift focus from merely utilizing theme
information as a style change signal to the actual writing style of the article itself.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        In the realm of text style analysis, traditional methods customarily depend on manually extracted
features such as word frequency, part-of-speech tags, and sentence length statistics, or utilize models
like TF-IDF [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], N-grams [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to capture patterns in text. Regrettably, these approaches frequently face
limitations due to the complexity of feature engineering and dependency on specific domains, making
their application across diverse text types or fields challenging.
      </p>
      <p>
        With the emergence of deep learning, models such as convolutional neural networks (CNN), recurrent
neural networks (RNN), long short-term memory networks (LSTM), and transformers have been
extensively utilized for text style analysis tasks. These models can autonomously learn intricate patterns
in text and efectively diferentiate between diferent text styles. For instance, RNN or LSTM’s ability to
process sequence data can capture context information in text [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], while transformers enhance the
model’s global information capture through self-attention mechanisms [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Despite significant advancements in text style analysis by deep learning models, they still confront
certain challenges. Firstly, models typically necessitate substantial labeled data for training, which
may be infeasible in some domains. Secondly, deep learning models often lack interpretability [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
making it dificult to comprehend how the model makes decisions. Lastly, for the task of detecting style
changes in multi-author text, deep learning models need to simultaneously capture the writing styles
of diferent authors and distinguish between them efectively.
      </p>
      <p>
        In recent years, contrastive learning [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], a novel learning method, has demonstrated robust
performance across various fields. Contrastive learning optimizes the model by maximizing similarity
between positive sample pairs and minimizing similarity between negative sample pairs, enabling the
model to learn more robust and generalizable feature representations. In the field of text style analysis,
contrastive learning is also employed to optimize the representation of text embeddings to boost the
model’s performance in style classification tasks [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        Previously, in style change detection tasks, Chen et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], utilizing a pretrained RoBERTa [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]
model achieved promising results. To address the "Collapse" [12] issue of traditional BERT models in
text semantic matching, a computational approach using contrastive learning fine-tuned the encoder,
computing cosine distance between sentences to learn improved sentence embedding representations.
This method reduces distances among similar sentences in vector space while increasing those between
dissimilar ones. By continuing this study’s findings, we further explored 2024’s highlighted topics and
text styles on new datasets.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Task and Datasets</title>
      <p>PAN 24 furnishes tasks [13] with three diverse complexity tiers. It necessitates identifying all positional
alterations in the stylistic conventions at the paragraphed level within an assigned text (i.e., scrutinize
for stylistic discrepancies between two successive paras). The primary distinction of these tasks lies in
the range of document subject matters:
• Easy: The paragraphs of a document cover a variety of topics, allowing methods to make use of
topic information to detect authorship changes.
• Medium: The topical variety in a document is small (though still present) forcing the methods
to focus more on style to efectively solve the detection task.</p>
      <p>• Hard: All paragraphs in a document are on the same topic.</p>
      <p>
        All documents are provided in English and may contain an arbitrary number of style changes.
However, style changes may only occur between paragraphs (i.e., a single paragraph is always authored
by a single author and contains no style changes). Each input problem is referenced by an ID (i.e., the
document that detects style changes), which is then used to identify the solution submitted for the
input problem. The ground truth data includes the number of authors and the binary labels of each pair
of consecutive paragraphs (1 for style changes, otherwise 0), but does not provide specific paragraph
author information [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Method</title>
      <p>
        Within this study, the contrastive learning method based on cosine phrase is used to train an encoder [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
The objective is to position the encoded sentences of identical authors closer in space and those derived
from diverse texts further apart. Post a certain level of training completion, we generate corresponding
classification labels by connecting a fully connected layer classifier.
      </p>
      <p>Figure 1 displays the model’s comprehensive framework, where pairs of paragraphs undergo identical
pooling strategy encoders and generate vector representations in the space. We’ve completed the text
embedding phase with writing style incorporated. Concurrently, a feature matrix (u, v, |u – v|) is used
to through contrastive analysis. This enables final classification via a FCNN classifier.</p>
      <p>Encoder</p>
      <p>Pooling
(u,v,|u-v|)</p>
      <p>FCNN</p>
      <p>Classifier
v
Paragraph 1</p>
      <p>Paragraph 2</p>
      <sec id="sec-4-1">
        <title>4.1. Encoder Training</title>
        <p>During the training phase of an encoder, each positive and negative instance pair is fed into the
encoder for embedding. Ideally, the cosine distance between positive pairs should be less than that
between negative pairs. That is, for any positive instances pair (, ) ∈ Ω pos and negative instances pair
(, ) ∈ Ω neg, there are:</p>
        <p>
          cos(i, j) &gt; cos(k, l)
Where x represents the embedding representation of the paragraph . The work of Su et al., [14] [15]and
Sun et al. [16] suggests an efective solution to such problems, here is the equation [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]:
 = log(1 +
        </p>
        <p>∑︁
(,)∈Ωpos,(,)∈Ωneg
 (cos(k,l)− cos(i,j)))
Where  &gt; 0 is a hyperparameter, which is taken as 20 in this experiment. The above equation is
used to optimize the encoder, and the cosine distance of the encoder output instances is evaluated
for correlation with the labels using the spearman metric, which assesses how well the relationship
between two variables can be described using a monotonic function.
(1)
(2)
Posi ve Pairs</p>
        <p>Nega ve Pairs
Embedding
Encoder</p>
        <p>(a)
Author 1</p>
        <p>Author 1
Posi ve Instances</p>
        <p>Author 1</p>
        <p>Author 2
Nega ve Instances</p>
        <p>S1
S3</p>
        <p>Cluster
(b)</p>
        <p>S2</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Classifier Training</title>
        <p>Upon coding-encoder training completion, we freeze the parameters of the encoder, encode instances of
the predefined paragraph pairs, extract the last layer of the model. We use the vectorial representations</p>
        <sec id="sec-4-2-1">
          <title>Author 1</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Author 2</title>
        </sec>
        <sec id="sec-4-2-3">
          <title>Author 2</title>
        </sec>
        <sec id="sec-4-2-4">
          <title>Author 3</title>
        </sec>
        <sec id="sec-4-2-5">
          <title>Author 3</title>
          <p>Example Document
He's at my place half the time and his fiance(e)'s
place the other half of the time. He's been
(homeless) couch surfing for several years and
only recently got engaged to his other partner.</p>
          <p>To the extent that they are there, it is with your
consent. The state has passed laws making sure
that vulnerable people (not saying he‘s one)
don’t get abused (not saying you‘re abusing him) Group 2:
Hostile territory, essentially. Mines, shelling,
drones, just angry Ukrainians with guns
(assuming these kids have no white flag above
their heads, but even then, its not always…
Lack of food and heating will not help them
either. And if they are unlucky, they will have
tshqeumadfsronmeaerbscyafpoirnmge.dEitshpeerctifhiceayl'llyl btoe…prevent Group 3:
He was and still is self directing in the eyes of the
US systems, but it would be easy to show that
based on his memory test scores here and in
Denmark that he was well beyond the point…</p>
        </sec>
        <sec id="sec-4-2-6">
          <title>Group 1:</title>
          <p>
            of these two matrices as paragraph embeddings (u, v), subtract them and take their absolute value, then
concatenate them with the original matrix to form a feature matrix (u, v, |u - v|) [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. This is fed into a
linear layer activated by tanh for classification. The prediction results are optimized using cross-entropy
loss and evaluated via F1-scores.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments and Result</title>
      <p>
        Firstly, it is imperative that we segment the oficial dataset to transform it into positive or negative
instances corresponding to each paragraph. In the oficial data set, there are no definitive authorship
information for each paragraph. This implies that if the sample transformation operation is based solely
on a dataset containing binary labels, both the size and accuracy of the training set will be insuficient.
(For instance, if two paragraphs written by the same author are mislabelled as negative instances while
others are deemed as positive instances, that may confuses the model.) To circumvent this issue, we
will employ the following strategy to generate positive and negative examples [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]:
• First, divide the paragraphs whose style has not changed into the same group, based on the labels.
• If the number of paragraphs in a group is greater than one, combine each of them in two to obtain
a positive instance.
      </p>
      <p>• Two-by-two combinations of negative instances between two adjacent but diferent groups.</p>
      <p>Through this method shown in Figure 3, a significant volume of high-quality instance pairs can be
obtained. Our method is exclusively utilized for the training set. For the validation set, we employ
positive and negative instance pairs transformed from original labels only.</p>
      <p>Secondly, we select the RoBERTa-base model as our pre-trained model and fed the prepared data to
it. Our hyperparameters are set as follows: For the encoder, the batch size is set to 24, the maximum
sequence length is 512, and the excess will be truncated. The initial learning rate is set to 1e-5, and
trained in 10 epochs; For classifiers, the batch size is set to 64, the initial learning rate is set to 5e-5, and
trained in 10 epochs.</p>
      <p>The optimized result with the best F1 score of 0.8074 was accomplished by RoBERTa as pre-trained
encoder model. The code and model were finally packaged in a docker file and submitted to Tira</p>
      <sec id="sec-5-1">
        <title>Positive Instances:</title>
      </sec>
      <sec id="sec-5-2">
        <title>Negative Instances:</title>
      </sec>
      <sec id="sec-5-3">
        <title>Para1</title>
      </sec>
      <sec id="sec-5-4">
        <title>Para2</title>
      </sec>
      <sec id="sec-5-5">
        <title>Para3</title>
      </sec>
      <sec id="sec-5-6">
        <title>Para4</title>
      </sec>
      <sec id="sec-5-7">
        <title>Para5</title>
        <p>platform for execution, leading to the final measure of the model’s performance. Table 1 provides the
scores obtained by our model in the oficial test set.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgements</title>
      <p>This work is supported by the Social Science Foundation of Guangdong Province, China (No. GD24CZY02).
[12] Y. Yan, R. Li, S. Wang, F. Zhang, W. Wu, W. Xu, Consert: A contrastive framework for
selfsupervised sentence representation transfer, arXiv preprint arXiv:2105.11741 (2021).
[13] E. Zangerle, M. Mayerl, M. Potthast, B. Stein, Pan24 multi-author writing style analysis (2024).</p>
      <p>URL: https://zenodo.org/records/10677876.
[14] X. Huang, H. Peng, D. Zou, Z. Liu, J. Li, K. Liu, J. Wu, J. Su, P. S. Yu, Cosent: Consistent sentence
embedding via similarity ranking, IEEE/ACM Transactions on Audio, Speech, and Language
Processing 32 (2024) 2800–2813. doi:10.1109/TASLP.2024.3402087.
[15] J. Su, M. Zhu, A. Murtadha, S. Pan, B. Wen, Y. Liu, Zlpr: A novel loss for multi-label classification,
arXiv preprint arXiv:2208.02955 (2022).
[16] Y. Sun, C. Cheng, Y. Zhang, C. Zhang, L. Zheng, Z. Wang, Y. Wei, Circle loss: A unified perspective
of pair similarity optimization, in: Proceedings of the IEEE/CVF conference on computer vision
and pattern recognition, 2020, pp. 6398–6407.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. B.</given-names>
            <surname>Casals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Elnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Freitag</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Korenčić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Smirnova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Taulé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ustalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle,
          <article-title>Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification, in: Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Q. A.</given-names>
            <surname>Bui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Visani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Prum</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.-M. Ogier</surname>
          </string-name>
          ,
          <article-title>Writer identification using tf-idf for cursive handwritten word recognition</article-title>
          ,
          <source>in: 2011 International Conference on Document Analysis and Recognition</source>
          , IEEE,
          <year>2011</year>
          , pp.
          <fpage>844</fpage>
          -
          <lpage>848</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ríos-Toledo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P. F.</given-names>
            <surname>Posadas-Durán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sidorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Castro-Sánchez</surname>
          </string-name>
          ,
          <article-title>Detection of changes in literary writing style using n-grams as style markers and supervised machine learning</article-title>
          ,
          <source>Plos one 17</source>
          (
          <year>2022</year>
          )
          <article-title>e0267590</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. X.</given-names>
            <surname>Huang</surname>
          </string-name>
          , L. He,
          <article-title>Ca-rnn: using context-aligned recurrent neural networks for modeling sentence similarity</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>32</volume>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Santhanam</surname>
          </string-name>
          ,
          <article-title>Context based text-generation using lstm networks</article-title>
          , arXiv preprint arXiv:
          <year>2005</year>
          .
          <volume>00048</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>Self-attention attribution: Interpreting information interactions inside transformer</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>35</volume>
          ,
          <year>2021</year>
          , pp.
          <fpage>12963</fpage>
          -
          <lpage>12971</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L. H.</given-names>
            <surname>Gilpin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. Z.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bajwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Specter</surname>
          </string-name>
          , L. Kagal,
          <article-title>Explaining explanations: An overview of interpretability of machine learning</article-title>
          ,
          <source>in: 2018 IEEE 5th International Conference on data science and advanced analytics (DSAA)</source>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>80</fpage>
          -
          <lpage>89</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Le-Khac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Healy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Smeaton</surname>
          </string-name>
          ,
          <article-title>Contrastive representation learning: A framework and review</article-title>
          ,
          <source>Ieee Access</source>
          <volume>8</volume>
          (
          <year>2020</year>
          )
          <fpage>193907</fpage>
          -
          <lpage>193934</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Puri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J</given-names>
            .
            <surname>Tworek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tezak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          , et al.,
          <article-title>Text and code embeddings by contrastive pre-training</article-title>
          ,
          <source>arXiv preprint arXiv:2201.10005</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <article-title>A writing style embedding based on contrastive learning for multi-author writing style analysis</article-title>
          .,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>2562</fpage>
          -
          <lpage>2567</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , W. Chen, Deberta:
          <article-title>Decoding-enhanced bert with disentangled attention</article-title>
          , arXiv preprint arXiv:
          <year>2006</year>
          .
          <volume>03654</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>