<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Aarhus, Denmark
£ schoech@uni-trier.de(C. Schöch); dudar@uni-trier.d(eJ. Dudar);fileva@uni-trier.de(E. Fileva);
sela@ucl.cas.cz(A. Šeļa)
ç https://christof-schoech.de/en(C. Schöch); https://artjomsh.github.io/web/(A. Šeļa)
ȉ</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Multilingual Stylometry: The Influence of Language on the Performance of Authorship Attribution using Corpora from the European Literary Text Collection (ELTeC)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>ChristofSchöch</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>JuliaDudar</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Evgeniia Fileva</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ArtjomsŠeļa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Czech Literature, Czech Academy of Sciences</institution>
          ,
          <addr-line>Prague</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Polish Language, Polish Academy of Sciences</institution>
          ,
          <addr-line>Kraków</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Trier Center for Digital Humanities, Trier University</institution>
          ,
          <addr-line>Trier</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Stylometric authorship attribution is concerned with the task of assigning texts of unknown, pseudonymous or disputed authorship to their most likely author, often based on a comparison of the frequency of a selected set of features that represent the texts. The parameters of the analysis, such as feature selection and the choice of similarity measure or classification algorithm, have received significant attention in the past. Two additional key factors for the performance and reliability of stylometric methods, however, have so far received less attention, namely corpus composition and corpus language. As a first step, the aim of this study is to investigate the influence of language on the performance of stylometric authorship attribution. We address this question using four diferent corpora derived fromEtuhreopean Literary Text Collection (ELTeC). We use machine-translation to obtain each corpus in the other three languages. We find that, as expected, the attribution accuracy varies between language-based corpora, and that translated corpora, on average, display a lower attribution accuracy compared to their counterparts in the original language. Overall, our study contributes to a better understanding of stylometric methods of authorship attribution.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Stylometry</kwd>
        <kwd>Authorship Attribution</kwd>
        <kwd>Multilingualism</kwd>
        <kwd>ELTeC</kwd>
        <kwd>Translation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Stylometric authorship attribution is the task of assigning texts of unknown, pseudonymous or
disputed authorship to their most likely author, often based on a comparison of the frequency
of a selected set of features that represent texts14[
        <xref ref-type="bibr" rid="ref20 ref23 ref5">, 26, 6, 23</xref>
        ].
      </p>
      <p>
        Traditionally, the way stylometric methods approach authorship attribution is to use the
frequencies of a large number of simple features, such as words, lemmas or character sequences,
for a determination of the degree of similarity between texts. These similarities, in turn, are
interpreted as an indicator of the likelihood for two texts to have been written by the same author:
the more similar the feature vectors, the more likely is identical authorship. The parameters of
the analysis, such as feature selection and the choice of similarity measure or classification
algorithm, have received significant attention in the past (e.g. [
        <xref ref-type="bibr" rid="ref1 ref22 ref9">1, 25, 27, 11</xref>
        ]). Two additional key
factors for the performance and reliability of stylometric methods, however, are corpus
composition and corpus language. They are relevant not only for the results in a specific case, but also
for the overall performance and reliability of stylometric methods of authorship attribution.
      </p>
      <p>Therefore, the aim of ongoing research by our group is to disentangle the influence of corpus
composition and language on the performance of stylometric authorship attribution: To what
extent do the attribution accuracy and robustness of such approaches depend on the language
of the materials? To what extent do they depend on the corpus composition, which can present
more or less challenging constellations of authorship and style? How do these two factors
interact with each other, and how do they interact with feature selection?</p>
      <p>This paper tackles one part of this issue, that of language, by investigating four distinct
but broadly comparable corpora in a classification scenario. The corpora are all derived from
the European Literary Text Collection (ELTeC) and contain at least three novels each by 8-10
diferent undisputed authors. These corpora can therefore be used as benchmarking datasets,
where the true authorship is known for all novels under investigation. Specifically, corpora
in English, French, Hungarian and Ukrainian are included. While the corpora contain similar
texts (fictional narrative prose from a similar time period, i.e., 1840–1920), they difer both in
terms of their exact composition, which can afect overall difÏculty of attribution and therefore
attribution performance, and in terms of their language, which can likewise afect attribution
performance.</p>
      <p>In order to investigate the role of language independently of corpus composition, all four
corpora were automatically translated into the other three languages using the DeepL machine
translation system (Pro version, used in the time period July to September 2023). This allows
us to vary language while keeping corpus composition stable, and in this manner, tease out
efects of both in terms of attribution performance (measured as classification accuracy).</p>
      <p>In addition, and in order not to disadvantage any one corpus or language by selecting
features that may not be suitable to them, a number of further parameters have been varied,
namely: the type of feature considered (word forms, lemmas, part of speech or characters);
the length of the feature sequence considered as the unit of analysis (unigrams, bigrams,
3grams, 4-grams or 5-grams); the total number of diferent features considered (from 50 to 2000
in several increments); and the length of contiguous textual segments considered (from 5000
words to 10000 words as well as using entire novels).</p>
      <p>Testing all of these various parameters, corpora and languages in all of their possible
combinations results in a large number of individual results. In order to make them accessible to
inspection and analysis, we have developed an interactive visualization that displays the
attribution quality (i.e. the classification accuracy) in a heatmap as a function of the parameters
described above. While the heatmap displays the accuracy for the entire range of features
and segment lengths at the same time, other parameters can be selected by the users in order
for the heatmap to be updated with the corresponding accuracy values. Two such heatmaps
are provided next to each other to enable convenient comparison of results between any two
configurations of parameters, whether corpora, languages or other settings.</p>
      <p>The present paper can be understood as a background publication to these heatmaps
published as a publicly available, interactive online showcase that is availablhettapts://showca
ses.clsinfra.io/stylometr.yThe target audience of this showcase are scholars and students
interested in a key methodological aspect of stylometry-based authorship attribution as well
as those interested in cross-lingual approaches within the digital humanities. The showcase is
intended both as a convenient way to document a specific research output and as a pedagogical
tool that could be used in workshop or classroom settings alike. Readers are encouraged to use
the showcase online, where much more detailed results can be explored, beyond the high-level
summary of results we present here.</p>
      <p>The detailed results are discussed below, and contribute to our general understanding of the
influence of language and corpus composition on stylometric results. Broadly speaking, we can
show that corpora of diferent language and composition lead to diferent attribution accuracy
levels and diferent best-performing features, an expected result. We can also show that
translated corpora (at least when all texts have been translated by the same machine translation
system) usually lead to a lower attribution accuracy, overall, compared to their counterpart in
the original language.</p>
      <p>In the following, we first discuss relevant prior work evaluating stylometric performance in
the context of multilingual corporsaec(tion 2). We then outline our research questions and
objectives (section 3), before describing our data and methodse(ction 4 and section 5). We
discuss the main outcomes with respect to attribution performance, detailing corpus-based
and language-based variations in resultse(ction 6). By way of a conclusion, we discuss the
consequences of our findings, document some limitations of this research and outline ongoing
research into the second aspect of this investigation, namely corpus composition.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Relevant prior work</title>
      <p>There has been a significant body of work investigating the methodological and theoretical
foundations of stylometric authorship attribution. Generally speaking, it seems fair to say
that most attention has been devoted to feature selection and the choice of distance measures
or classification algorithms. While much research is focused on English-language materials,
a limited amount of attention has been given to the question of variation across languages.
Only rarely, however, has the influence of corpus composition on attribution performance been
assessed beyond general statements about the advantages of generically-homogeneous corpora.
In the following, we discuss some of this research in more detail.</p>
      <p>
        With respect to investigations of feature selection, John Burrow5s, 4[] proposes multiple
measures each focused on a diferent part of the frequency spectrum: Delta, Iota and Zeta.
Smith and Aldridge2[
        <xref ref-type="bibr" rid="ref4">5</xref>
        ] use word forms in order to focus on the best setting for most frequent
words and word vector onset and have found that “a word frequency vector of between 200 and
300 words gives the most accurate results” when using Burrows’ Delta. They also show that
shifting the feature vector onset, efectively skipping the most frequent features, is detrimental
to attribution accuracy.
      </p>
      <p>
        In addition to varying the number of most frequent words, examples of using features other
than words can also be found in the research literature. In particular, Stamata2t7o]sf[ocuses
on the efectiveness of using character n-grams. He has conducted tests of 3-grams and
frequent words on texts of diferent genres and topics. The researcher views as an advantage
of n-grams that they capture linguistic patterns and may therefore be suitable to represent
an author’s style beyond word choice. He has obtained the best results with a “model based
on character 3-grams in combination with an SVM classifier with linear kernel”, and suggests
using this as a baseline for authorship attribution. Building on the conventional use of word
frequencies, Halteren et al.1[
        <xref ref-type="bibr" rid="ref2">3</xref>
        ] investigate the application of syntactic features in stylometry.
Their research highlights how syntactic patterns, based on word class tags, complement lexical
features in finding what they call the “human stylome”, i.e. individual measurable linguistic
features that help distinguishing between authors. The study shows that although 100%
efifciency results could not be achieved, lexical features perform slightly more efÏciently than
syntactic features and that in general, a greater number and/or variety of features improves the
results. In a similar vein, Gorman12[] uses syntactic information for the attribution of short
texts.
      </p>
      <p>
        With respect to distance or similarity measures, Argamo1n] [proposes a geometric
interpretation of Burrows’ Delta and derived several alternatives to this particular distance measure,
among them Quadratic Delta. Smith and Aldridg2e5[] propose to move beyond distance-based
comparisons and suggest the adoption of the Cosine similarity instead, where the directions
of the vectors are compared without being influenced by their length. Evert et al1.1[] follow
this important new direction and propose Cosine Delta, which combines z-score normalization
with the Cosine measure. They have confirmed the finding by Smith and Aldridge [
        <xref ref-type="bibr" rid="ref22">25</xref>
        ], namely
that performance is more stable for longer feature vectors when Cosine Delta is used. In
addition, they show that this efect holds across several languages, at least for long fiction narratives
(novels). In several cases, interactions between various factors (in particular between feature
vector length, distance measure and language) have been observed.
      </p>
      <p>
        With respect to languages, many methodological investigations focus on English corpora
only, as is the case with Burrows5[] or Smith and Aldridge2[
        <xref ref-type="bibr" rid="ref4">5</xref>
        ]. Rybicki and Eder [
        <xref ref-type="bibr" rid="ref16">19</xref>
        ],
however, show that results vary widely as a function of language when they investigated feature
vector length and onset for multiple corpora in diferent languages and covering various
literary genres. Evert et al. 1[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] test all their hypotheses on English, German and French corpora
consisting of novels, with only minor diferences between languages. However, the extent to
which the diferences in performance can be explained by language, on the one hand, and by
the inevitable diferences between corpora with respect to their genre and/or composition, is
not within the scope of their research.
      </p>
      <p>
        Stylometric research into translated works does exist, but is relatively rare. Rybicki and
Heydel [
        <xref ref-type="bibr" rid="ref19">22</xref>
        ] show that in the case where several diferent translators translate the works of a
given author, their stylistic signature can be detected by stylometric methods. In other cases,
especially in corpora involving multiple authors and multiple translators, the authorial signal
appears to be stronger than the translator’s signa2l1[]. In a recent study, Rybicki [
        <xref ref-type="bibr" rid="ref17">20</xref>
        ] focuses
on network analysis, using an extensive collection of translated texts from diferent genres
and languages into Polish. He observes that the translated and original texts reveal stylistic
similarities, from which it can be concluded that the translations are an extension of native
literary traditions. Rybicki also notes that the distinctiveness of genres becomes visible during
stylometric analysis.
      </p>
      <p>There are also several relevant studies that use machine translated texts. This approach
underlies Cross Language Authorship Attribution, a concept introduced by Bogdanova and
Lazaridou [2]. Although the quality of automatically translated texts may distort an author’s
style, translation is one way of bringing texts written in diferent languages “into one space”.
Machine translation in this study is combined with a focus on lexical and higher-level features.
The best results were achieved by combining machine translation and k-Nearest Neighbors.
Based on this study, Mikros and Boumparis17[] apply the machine translation method for
the language pair Greek-English. They conclude that when trained and tested on
machinetranslated texts, attribution accuracy is not strongly afected, compared to training and testing
on original texts, but that the relevant features difer between the two languages, making
crosslinguistic authorship attribution unreliable.</p>
      <p>
        As mentioned above, the influence of corpus composition on attribution performance has
so far not been investigated in any considerable detail. Smith and Aldrid2g5e] [observe that
a chronologically more diverse corpus leads to lower attribution accuracy but do not go into
much detail. Kestemont et al. 1[
        <xref ref-type="bibr" rid="ref4">5</xref>
        ] show that cross-genre attribution is more challenging than
within-genre attribution. A challenge for this kind of investigation is the availability of
meaningful, high-quality metadata to assess corpus composition.
      </p>
      <p>Based on this review of relevant research, we can conclude that we should expect some
variation in attribution accuracy depending on language. Also, there is a clear need to
investigate the role of corpus composition for the accuracy and robustness of stylometric authorship
attribution.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Research questions and objectives</title>
      <p>The overarching research question that ongoing work of our team addresses is to what
extent, how, and under what conditions, both language and corpus influence the performance of
stylometric methods of authorship attribution. The objective of our research, therefore, is to
provide detailed data evaluating the accuracy and robustness of stylometric methods for
corpora of diferent compositions and in diferent languages. As a first step, the present paper
aims to determine how the results of stylometry are influenced by the language of the texts,
when the corpus composition remains the same, that is, using corpora translated into multiple
languages. How does accuracy vary across languages? How does it change when a given
corpus is translated into a diferent language? What kind of interaction can we observe between
feature selection and language, as well as between sample size?</p>
      <p>In a second step, not covered by the present paper, we aim to determine how corpus
composition afects attribution accuracy. In order to address this question, a formalization of corpus
composition needs to be designed that quantifies the level of difÏculty, for an attribution task,
of a given corpus. How does such a measure need to be designed in order to adequately
represent corpus difÏculty in terms of authorship attribution? What types of metadata, for example
regarding subgenre, time period or formal aspects, are most strongly correlated to diferences
in attribution accuracy? Does a higher corpus difÏculty as measured based on metadata
correlate with lower attribution accuracy? What is the efect of translation on this relationship?
Our preliminary work on this issue is discussedsiunbsection 7.3 of this paper.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Data</title>
      <p>
        The dataset was derived from theEuropean Literary Text Collection (ELTeC; see [
        <xref ref-type="bibr" rid="ref2 ref21">24, 3</xref>
        ]) in order
to build corpora in four languages: French, English, Hungarian, and Ukrainian. The choice of
languages is determined by our desire to cover a number of diferent language groups:
Romance, Germanic, Finno-Ugric, and Slavic languages. In this way, one can illustrate the
efectiveness of the stylometric method and see how language afects the analysis results, for the
four language groups.
      </p>
      <p>The original ELTeC corpora each include one hundred novels. For our research, a selection of
novels was made, retaining novels by 8 to 10 authors, each represented with three novels from
ELTeC. This resulted in the following collections of texts: the English and the French corpora
include each 30 novels, the Hungarian corpus 27 novels and the Ukrainian corpus consists of
24 novels (see key information on the corpora iTnable 1).</p>
      <p>Each of the corpora was translated into the other three languages, thus the entire corpus
includes 16 sub-corpora. The dataset also includes a metadata table for each corpus, with
information about each of the novels. The metadata table includes information about the author,
year of publication of digital and print editions, language, number of words in the novel, as
well as subgenre (social, historical, adventure, detective or sentimental novel, bildungsroman
or other) and narrative perspective of the novel (heterodiegetic, homodiegetic, epistolary,
dialogue or mixed). This metadata was collected from experts in each of the relevant literary
traditions who based their information on a reading of both the novels and of relevant
secondary literature.</p>
      <p>Data preparation involved two stages: translation of texts into the other languages and their
linguistic annotation. The texts were translated automatically using DeepL Pro. For
subsequent analysis, extracting lemmas and POS (part of speech) tags was necessary, a task we
accomplished using the SpaCy library (version 3.7) for both original and translated texts.
Additionally, unigrams and n-grams (from 2 to 5) were extracted using the stylo packa1g0e].[</p>
    </sec>
    <sec id="sec-5">
      <title>5. Method(s)</title>
      <p>Fundamentally, we used an authorship attribution classification task with leave-one-out
crossvalidation. Inspired by Rybicki and Eder19[], we used grid search over the features space to
assess the general performance of authorship attribution and quantified the attribution
performance in terms of its accuracy (mean attribution accuracy score for each condition as well
as Cohen’s Kappa for each condition). The goal was not to optimize for performance, but to
cover the space of reasonable approaches that work for diferent languages and to understand
the nature of diferences in performance across languages and corpora.</p>
      <sec id="sec-5-1">
        <title>5.1. Features</title>
        <p>While in many scenarios, simple word form frequencies still prove efective, other kinds of
features can in some settings be advantageous, for example for short texts or agglutinative
languages [18]. Therefore, we did want to vary the feature types to some extent, in order
not to miss conditions which work well in one or the other of the less-frequently investigated
languages. There are three main levels of variation: types of features, sample size and feature
vector length.</p>
        <p>In terms of types of features, we used frequencies of word forms, lemmas, part-of-speech
(POS) tags and characters. Each feature was cut to n-grams of diferent size: 1-3 for words
and lemmata, 2-5 for character and POS n-grams (because the number of POS and character
unigrams is too limited to be useful). Each n-gram length was tested independently.</p>
        <p>With respect to sample size, frequency-based approaches to authorship attribution naturally
depend on the available size of the text. There is a considerable variation in text sizes in
ELTeC (as shown in Table 1); to mitigate this, we draw a random sample of consecutive tokens (a
“chunk”) for each text based on the shortest text across all corpora (10,000 words). We use test
sizes of 5,000 to 10,000 tokens for word-based features, and 10,000 to 50,000 for character-based
ones. The number of available tokens per text difers dramatically between words and
characters; so sample sizes mean diferent things for n-grams based on these features. To account for
the variability that is introduced with taking only one limited sample out of sometimes very
large novels, at each step we take a random consecutive sample out of all available ‘chunks’
and record performance 100 times. As the last step, we perform classification on full-length
texts.</p>
        <p>In terms of feature vector length, vectors that represent texts are constructed based on most
frequent feature length cutofs. We used 50 to 2000 features with incrementally increasing step
sizes.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Classification task</title>
        <p>For author-based text classification, we use the Support Vector Machine classifier and perform
leave-one-out cross-validation. This means that in each step, one text is removed from the data,
the model is then trained on the remaining texts, and then the authorship of the left-out text
is predicted and the result is recorded. This process continues until all texts have been left out
once. For each language combination, we run 100 iterations of leave-one-out cross validation
classification, taking a sample from each text (chunks of consecutive sequences of size n) at
random. Additionally, we run a single full-text analysis for each corpus.</p>
        <p>We report two performance measures: simple accuracy (proportion of correctly predicted
authors) and Cohen’s kappa (typically used for measuring inter-annotator agreeme7n])t. [The
latter measure, while highly correlated to accuracy, allows us to partly ofset the diferent
amount of classes across corpora, since it provides an inter-rater agreement score adjusted for
random classification.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Visualization</title>
        <p>To create the interactive visualizations of the showcase, we used Bokeh, a Python library
offering extensive capabilities in crafting interactive and dynamic data visualizations. The
visualization – which is available online ahtttps://showcases.clsinfra.io/stylometr–y consists of
two heatmaps, enabling users to compare and contrast two sets of results simultaneously. In
this article, we focus on the results for the Hungarian corpus, which are
presentesduibnsection 6.2. For results from other corpora, please seection 8 (appendix) and/or visit the online
showcase.</p>
        <p>The x-axis represents various settings of the most frequent features (MFF) used in a specific
analysis, while the y-axis denotes distinct sample sizes, ranging from shorter text snippets to
entire novels (”full novel”). Words and characters are treated diferently and analyzed across
diverse sample sizes (seesubsection 5.1 above). Each heatmap cell correlates MFF and
sample size, with the color intensity indicating accuracy levels, ofering a visual approach to the
accuracy obtained in the analysis. A mouseover provides further information, such as the
features used as well as numerical indications of accuracy and Cohen’s Kappa. When there is no
data corresponding to selected feature combinations, the plots display a ”No data available for
these selectors” message. The range of MFF values depends on the features; in some cases, as
for character and POS bigrams, the theoretical maximum of features is limited.</p>
        <p>Users can engage with the data through several selectors, enhancing the exploration and
analysis process. The available choices include:
1. Corpus selector: Enables the selection from a range of available corpora, facilitating
comparative studies;
2. Feature level selector: Permits users to toggle between ”words” and ”chars” (characters)
feature types;
3. Feature type selector: Allows for the choice between plain text, lemmas, or POS tags;
4. Ngram size selector: Ofers options to choose n-gram sizes from 1 to 5, allowing for
detailed linguistic patterns examination.</p>
        <p>A color scale on the right side of each graph delineates the accuracy levels, ranging from 0 to
1. Here, cooler tones like greens, blues and purples signify lower values, whereas warmer hues,
like red and orange, indicate higher values. This color coding serves as a primary indicator of
stylometric efectiveness across various parameter settings, vividly illustrating the success rate
of stylometric analysis as a function of the chosen corpus and the selected criteria.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>In relation to the main outcomes, we aim to present a concise overview of some of our key
ifndings. First, we focus on evaluating the overall performance of authorship attribution, based
on mean accuracy scores across multiple settings, for all corpora. Second, we provide a detailed
description of the main outcomes for the Hungarian corpus and its translations. We focus on
one language for reasons of space, but readers can find a detailed description of the performance
of the other three languages in the appendixse(ction 8) to this paper.</p>
      <sec id="sec-6-1">
        <title>6.1. Attribution performance across all corpora</title>
        <p>Regarding overall performance, we discovered several parameters that consistently yield the
best results across all corpora and translations. At the word level, the following parameters
showed the highest scores: unigram for both feature types (plain word forms and lemmas),
with only minimal diferences. Meanwhile, at the character level, very high performance was
observed using 5-grams based on both plain word forms and lemmas as feature types and using
full novels as sample size.</p>
        <p>In Figure 1and Figure 2, the average classification accuracy at the word level is illustrated,
employing unigram plain word forms, with sample sizes ranging from 7,000 to 9,000 and from
10,000 to full novel size, respectively. As observed in the figures, the clear winner among the
original texts is the Hungarian corpus. Translations from Hungarian into three other languages
also demonstrate very high performance, slightly decreasing to 0.6 and 0.7 from 0.86 compared
to original texts when full texts and sample size of 10000 are taken into account. Concerning
smaller sample sizes, the decrease is more noticeable, achieving an accuracy of 0.4-0.5, while
the accuracy of original texts is 0.8. The second-best performance comes from the original
Ukrainian corpus; however, translations from Ukrainian have low performance, falling to 0.5
and 0.4 from the original scores of 0.76 and 0.65 for large and smaller sample sizes respectively,
indicating that translations from Ukrainian are particularly challenging for attribution tasks.
The best performance among translations from Ukrainian, surprisingly, is shown by the
Hungarian language.</p>
        <p>The French and English corpora have almost similar performance when the original texts are
analyzed. However, translations from English at the word level perform lower compared to the
original texts, and the diference in performance is more pronounced with smaller sample sizes,
falling from 0.5 to 0.3-0.4. On the other hand, at the character level, translations into Hungarian
outperform original texts with the following settings: 5-gram based on both lemmas and plain
word forms.</p>
        <p>Remarkably, the smallest diference in performance between translated texts and the
original is observed in the French corpus, where the diference is smaller than 0.1 for large sample
sizes and only slightly greater than 0.1 for smaller sample sizes. It appears that the authors’
style is best preserved in translations from French. All three translations from French
demonstrate similar performance; however, translations into English slightly outperform the other
translations.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Hungarian corpus attribution performance</title>
        <p>As already discussed earlier, the Hungarian corpFuisg(ure 3) demonstrates the highest accuracy
scores overall, reaching approximately 80–95% across many sample sizes; starting with samples
of 8000 words, results are reliably very accurate. This achievement is realized by employing
the following settings: the number (N) of the most frequent words ranges from 50 to 700, with
the analysis based on unigram plain word forms. Although slightly less efective, classification
based on unigram lemmas still produces high scores. Regarding the analysis using character
n-grams as features, the performance remains remarkably high, fluctuating between 70-93% for
sample sizes starting from 4000, with the most remarkable performance observed when 5-gram
character sequences are used with full novelFsig(ure 4). In this scenario, the attribution based
on characters, derived from word forms with the number of MFF from 50 to 500, maintains a
constant accuracy rate of over 90% for full novels.</p>
        <p>When evaluating the accuracy of translations from Hungarian into other languages, it
becomes evident that the performance is noticeably lower compared to original novels. One of
the few configurations displaying high accuracy (over 90%) is observed in translations into
Ukrainian utilizing 5-gram characters based on plain word forms. For lemmas, the Ukrainian
translation, in some cases, outperforms the original texFtsig(ure 5). As for the attribution
at the word level, high accuracy is observable only for full novel sample sizes. Translations
into English and Ukrainian exhibit very similar performance, whereas French displays lower
accuracy.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <sec id="sec-7-1">
        <title>7.1. Findings</title>
        <p>In conclusion, we would like to summarize some of the key findings of this study, but also
discuss a number of limitations and next steps for research in this area.</p>
        <p>We can discern consistent patterns across all languages and translations when analyzing their
performance. These trends remain mostly consistent whether we are examining analyses based
on words, lemmas or character n-grams.</p>
        <p>Optimal performance consistently emerges when the analysis is based on entire novels
rather than shorter samples. When using samples, results tend to improve as the size of the
analyzed samples increases. This finding is consistent with previous studies4[, 8, 9].</p>
        <p>
          Conversely, increasing the number of features used for attribution beyond a certain value
tends to lead to a deterioration in classification scores in most cases. The decline in
performance becomes noticeable at full novels and word unigrams beyond roughly 300-800 most
frequent words for French, English, and Hungarian (but this is not the case for Ukrainian).
Furthermore, the impact is considerably less pronounced for word bigrams and trigrams. This
ifnding is consistent with results from distance-based attribution studies when using Burrows’
Delta, but not when using Cosine or Cosine Delta25[
          <xref ref-type="bibr" rid="ref9">, 11</xref>
          ]. This phenomenon may be caused
either by an increasing proportion of features driven by topic rather than authorship, or by an
increased amount of noise in the signal, when using higher numbers of features.
        </p>
        <p>Performance for corpora is always better when they are analyzed in their original language,
rather than in a translation into another language. However, this might be due to the fact that
we have used a unique translator, as it were, this fact being a major limitation of our study (see
below).</p>
      </sec>
      <sec id="sec-7-2">
        <title>7.2. Limitations</title>
        <p>
          The finding, mentioned above, that the translated corpora always perform less well, in our
experiments, than the original corpora, may very well be an artefact of our method of translating
these texts. Given that we used DeepL for all translations, the translations could be said to have
all been produced by the same ‘agent’, leading to a certain homogenization of their style and
lexicon, which we would expect to be a challenge for stylometric authorship attribution. In
contrast, translations of literary texts are usually performed by a range of diferent translators,
some of which maintain privileged relationships with certain authors, in a constellation where
authorial and translatorial signal may reinforce each other. At the same time, the question of
the (in)visibility of the translators in stylometry has been addressed several tim21e,s1[
          <xref ref-type="bibr" rid="ref5">6</xref>
          ].
        </p>
        <p>Another limitation with respect to the corpora is that for each language, we have only one
particular corpus available. This means that, given how important the genre of a corpus as
well as the corpus composition are likely to be, we cannot generalize our results based on
these particular corpora to the languages in general.</p>
      </sec>
      <sec id="sec-7-3">
        <title>7.3. Next steps</title>
        <p>As mentioned above, the research presented here only covers one of two aspects that we
consider relevant to our overall research question. Our corpora difer between each other both in
language and in composition. Therefore, the investigation of the influence of language and of
translation described here needs to be complemented by an analogous investigation into the
influence of corpus composition on the results.</p>
        <p>We have already made significant progress in this investigation, notably by collecting
relevant metadata about all four corpora and by computationally modeling the relationship
between corpus composition and attribution performance. With respect to the metadata, we have
collected information about the year of publication, the narrative form and the subgenre for
each novel in each corpus. To enable cross-corpus application of the subgenre labels, they come
from a small, closed set of eight diferent, relatively broad categories, including one category
called ’other’. With respect to the computational model of the relation between corpus
composition and attribution accuracy, we have fitted a multiple generalized Bayesian linear model
that takes into account within-author and between-author variability and allows us to assess
the positive or negative influence of metadata diferences on textual distinctiveness (and,
therefore, stylometric performance). However, further work on this issue is needed and presenting
this work warrants a separate publication.</p>
        <p>Apart from this, our research opens up several avenues for future research, in order to
attempt to close some of the gaps just described in the section on limitations. For instance, the
same kind of analysis could be performed using several corpora of diferent genres, size and
composition for each language, in order to assess the degree of within-language variation and
the robustness of diferences between languages that we have observed in this study. In
addition, a study could be conceived that varies corpus composition deliberately and systematically,
based on a significantly larger corpus with document-level metadata (at least) on subgenre and
narrative perspective.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>8. Data and code</title>
      <p>Data and code are available online ahtttps://gitlab.clsinfra.io/cls-infra/d3(3DOI: https:
//doi.org/10.5281/zenodo.1108020)5.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments References</title>
      <p>The research presented here as been performed as part of CLS INFRA, a project supported by
funding from the European Union’s Horizon 2020 research and innovation programme under
grant agreement No 101004984.</p>
      <p>D. Bogdanova and A. Lazaridou. “Cross-Language Authorship Attribution”.
IPnr:oceedings of the Ninth International Conference on Language Resources and Evaluation
(LREC’14). Reykjavik: Elra, 2014, pp. 2015–2020. url:http : / / www . lrec - conf . org / pro
ceedings/lrec2014/pdf/145%5C%5FPaper.pd. f
[18]
[27] E. Stamatatos. “On the robustness of authorship attribution based on character n-gram
features”. In:Journal of Law and Policy 21.2 (2013). url: https://brooklynworks.brookla
w.edu/jlp/vol21/iss2/7./</p>
    </sec>
    <sec id="sec-10">
      <title>Appendix</title>
      <p>Note that all result data can be inspected online in our interactive visualizatiohnttapts://show
cases.clsinfra.io/stylometr. yA selection of findings is documented here for convenience.</p>
    </sec>
    <sec id="sec-11">
      <title>A. Ukrainian</title>
      <p>The Ukrainian corpus demonstrates lower performance compared to the Hungarian original
corpora, achieving accuracies ranging from 50-80% when using unigram lemmas. Analyses
based on word forms yields higher scoreFsig(ure 6, 60-85%), particularly when the number of
features is not higher than 400-500. An interesting observation is that variation in accuracy
is a bit less across parameters in the Ukrainian original texts than for other corpora. When
evaluating the accuracy of character-level analysis, high scores (70-85%) are observed using
3-5-gram characters with full novel sample size in word forms and lemmas settiFniggsu(re 7).</p>
      <p>Translations of the Ukrainian corpus into other languages demonstrate significantly lower
results than the original texts, with accuracies surpassing 50% only for full novel sizes in
wordlevel analysis with unigram word forms and lemmas. Among the translations from Ukrainian,
the English translation shows the highest results, reaching up to 75% for full novel sample
size (Figure 8). This observation holds true for analyses based on full novel sample size using
character n-grams on translations as well. In most cases, the performance of translation into
English achieves 80% accuracy for 5-gram characters in both word forms and lemmas.
Concerning the mean performance, the translation into Hungarian has the highest mean accuracy
in most settings.</p>
    </sec>
    <sec id="sec-12">
      <title>B. French</title>
      <p>The French corpus shows a bit lower average performance compared to Ukrainian and
Hungarian, with accuracy ranging from 40-70%, and in some cases, up to 80-95% for original texts based
on unigram word forms and lemmas with N features up to 550 and a sample size starting from
8000 (Figure 9). Notably, the mean accuracy of original texts based on unigram word forms
with sample size of 10000 and full novels is very high and reaches 83%, which approximates
the performance of the Hungarian original corpus.</p>
      <p>When discussing character n-gram features, we observe high scores (up to 80-95%)
exclusively for full novels, with n-gram sizes ranging from 3F-i5gu(re 10). This observation holds
true for both lemmas and plain word forms.</p>
      <p>Translations from French show a moderate decrease in performance compared to the original
texts in both lemma- and word-form-based analyses, maintaining high accuracy only for the
full novel sample size. Regarding the character n-grams, the scores of translations are almost
identical to those of the original novels, with high performance achieved only for the full novel
sample size.</p>
    </sec>
    <sec id="sec-13">
      <title>C. English</title>
      <p>The English corpus F(igure 11) yields the lowest performance among all corpora for original
texts, with accuracy occasionally reaching 93% only on the full novel sample size in unigram
lemma-based analysis at the word level. Apart from these instances, the corpus’s performance
displays high variability, including the lowest results around 20% for both lemma-based and
plain word form-based analyses at the word level.</p>
      <p>Regarding the classification performance when using character n-grams, we observe a
similar trend as seen in the case of the French original corpus, where high accuracy is evident only
for full novels. However, for the English corpus, the performance is slightly lower, reaching
86% in the best cases (Figure 12).</p>
      <p>Importantly, the Hungarian translation of the English corpuFisg(ure 13) approximates the
performance of the original novels, and in some cases even outperforms the original novels.
The highest accuracy of up to 96% (lemma-based analysis) and 90% (word forms-based) is
achieved for full novel sample size at the word level. However, the performance decreases
rapidly with an increase in the number of features and decrease in sample sizes.</p>
      <p>The translation into UkrainianF(igure 14) also demonstrates high performance, up to 86%
for full novels, although the average accuracy is lower compared to the original corpus and the
translation into HungarianF(igure 13). On the other hand, the translation into French achieves
best performances at around 84% for lemma-based analysis, with most cases hovering around
20-30% (Figure 15). Regarding the analysis of translations at the character level, Hungarian
translation outperforms the original corpus, in full novels sample size setting, while French
translations exhibit lower accuracy even for full novels, reaching a score of 80% only for the
1500 and 2000 features setting for word forms and lemma-based analysis.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Argamon</surname>
          </string-name>
          . “
          <article-title>Interpreting Burrows's Delta: Geometric and Probabilistic Foundations”</article-title>
          .
          <source>In: Literary and Linguistic Computing 23.2</source>
          (
          <issue>2008</issue>
          ), pp.
          <fpage>131</fpage>
          -
          <lpage>147</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqn00 3.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Burnard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Odebrecht</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Schöch</surname>
          </string-name>
          . “In Search of Comity:
          <article-title>TEI for Distant Reading”</article-title>
          .
          <source>In: Journal of the Text Encoding Initiative</source>
          <volume>14</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .4000/jtei.3500.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Burrows</surname>
          </string-name>
          . “
          <article-title>All the Way Through: Testing for Authorship in Diferent Frequency Strata”</article-title>
          .
          <source>In: Literary and Linguistic Computing 22.1</source>
          (
          <issue>2007</issue>
          ), pp.
          <fpage>27</fpage>
          -
          <lpage>47</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqi067.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Burrows</surname>
          </string-name>
          . “
          <article-title>The Englishing of Juvenal: Computational Stylistics and Translated Texts”</article-title>
          .
          <source>In: Style 36.4</source>
          (
          <year>2002</year>
          ). url: https://www.jstor.
          <source>org/stable/10.5325/style.36.4.67.7</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [6]
          <string-name>
            <surname>J. Byszuk. “</surname>
          </string-name>
          <article-title>Analysis in authorship attribution”. InS:urvey of Methods in Computational Literary Studies</article-title>
          . Ed. by
          <string-name>
            <given-names>C.</given-names>
            <surname>Schöch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dudar</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E.</given-names>
            <surname>Fileva</surname>
          </string-name>
          . Trier: Cls Infra,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .5281/zenodo.7782363.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cohen</surname>
          </string-name>
          .
          <article-title>“A CoefÏcient of Agreement for Nominal Scales”</article-title>
          .
          <source>In: Educational and Psychological Measurement 20.1</source>
          (
          <issue>1960</issue>
          ), pp.
          <fpage>37</fpage>
          -
          <lpage>46</lpage>
          . doi:
          <volume>10</volume>
          .1177/001316446002000104.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Eder</surname>
          </string-name>
          . “
          <article-title>Does size matter? Authorship attribution, small samples</article-title>
          , big problem”D.Iign-:
          <source>ital Scholarship in the Humanities 30.2</source>
          (
          <issue>2015</issue>
          ), pp.
          <fpage>167</fpage>
          -
          <lpage>182</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqt06 6.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Eder</surname>
          </string-name>
          . “
          <article-title>Short Samples in Authorship Attribution: A New Approach”</article-title>
          .
          <source>BIno:ok of Abstracts of the Digital Humanities Conference</source>
          <year>2017</year>
          ).
          <source>Montréal: Adho</source>
          ,
          <year>2017</year>
          . url:https://dh 2017.adho.org/abstracts/341/341.pd.f M.
          <string-name>
            <surname>Eder</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rybicki</surname>
            , and
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kestemont</surname>
          </string-name>
          . “
          <article-title>Stylometry with R: A Package for Computational Text Analysis”</article-title>
          .
          <source>In:The R Journal 8.1</source>
          (
          <issue>2016</issue>
          ), p.
          <fpage>107</fpage>
          . doi:
          <volume>10</volume>
          .32614/rj-2016-007.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Evert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Proisl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Jannidis</surname>
          </string-name>
          , I. Reger,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pielström</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Schöch</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Vitt</surname>
          </string-name>
          . “
          <article-title>Understanding and explaining Delta measures for authorship attribution”</article-title>
          .
          <source>DInig:ital Scholarship in the Humanities 32.2</source>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .1093/llc/fqx02 3.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Gorman</surname>
          </string-name>
          . “
          <article-title>Universal Dependencies and Author Attribution of Short Texts with Syntax Alone”</article-title>
          .
          <source>In:Digital Humanities Quarterly 16.2</source>
          (
          <year>2022</year>
          ). url: https://www.digitalhumanitie s.org/dhq/vol/16/2/000606/000606.htm.l H. v. Halteren,
          <string-name>
            <given-names>H.</given-names>
            <surname>Baayen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Tweedie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Haverkort</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>A. Neijt. “</surname>
          </string-name>
          <article-title>New Machine Learning Methods Demonstrate the Existence of a Human Stylome”</article-title>
          .
          <source>InJ:ournal of Quantitative Linguistics 12.1</source>
          (
          <issue>2005</issue>
          ), pp.
          <fpage>65</fpage>
          -
          <lpage>77</lpage>
          . doi:
          <volume>10</volume>
          .1080/09296170500055350.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>D. I. Holmes.</surname>
          </string-name>
          “
          <article-title>The Evolution of Stylometry in Humanities Scholarship”</article-title>
          .
          <source>LIinte:rary and Linguistic Computing 13.3</source>
          (
          <issue>1998</issue>
          ), pp.
          <fpage>111</fpage>
          -
          <lpage>117</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/13.3.111.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Kestemont</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Luyckx</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Daelemans</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Crombez</surname>
          </string-name>
          . “
          <string-name>
            <surname>Cross-Genre Authorship Verification Using</surname>
          </string-name>
          <article-title>Unmasking”</article-title>
          .
          <source>In: English Studies 93.3</source>
          (
          <issue>2012</issue>
          ), pp.
          <fpage>340</fpage>
          -
          <lpage>356</lpage>
          . doi:
          <volume>10</volume>
          .1080 /0013838x.
          <year>2012</year>
          .
          <volume>668793</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lee</surname>
          </string-name>
          . “
          <article-title>Do language combinations afect translators' stylistic visibility in translated texts?”</article-title>
          <source>In: Digital Scholarship in the Humanities 33.3</source>
          (
          <issue>2018</issue>
          ), pp.
          <fpage>592</fpage>
          -
          <lpage>603</lpage>
          . doi:
          <volume>10</volume>
          .1093/ll c/fqx056.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>G.</given-names>
            <surname>Mikros</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Boumparis</surname>
          </string-name>
          . “
          <article-title>Cross-linguistic Authorship Attribution</article-title>
          and
          <string-name>
            <given-names>Author</given-names>
            <surname>Proifling</surname>
          </string-name>
          .
          <article-title>Machine translation as a method for bridging the language gap”</article-title>
          .
          <source>InD:igital Scholarship in the Humanities 39.3</source>
          (
          <issue>2024</issue>
          ), pp.
          <fpage>954</fpage>
          -
          <lpage>967</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqae02 8.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Nini</surname>
          </string-name>
          .
          <article-title>A Theory of Linguistic Individuality for Authorship Analysis</article-title>
          . Cambridge University Press,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .1017/9781108974851.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rybicki</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Eder</surname>
          </string-name>
          . “
          <article-title>Deeper Delta across genres and languages: do we really need the most frequent words?” InL:iterary</article-title>
          and
          <source>Linguistic Computing 26.3</source>
          (
          <issue>2011</issue>
          ), pp.
          <fpage>315</fpage>
          -
          <lpage>321</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqr03.1
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rybicki</surname>
          </string-name>
          .
          <article-title>“A Third Glance at a Stylometric Map of Native and Translated Literature in Polish”</article-title>
          . In:
          <article-title>Retracing the History of Literary Translation in Poland</article-title>
          . New York: Routledge,
          <year>2021</year>
          , pp.
          <fpage>247</fpage>
          -
          <lpage>261</lpage>
          . doi:
          <volume>10</volume>
          .4324/
          <fpage>9780429325366</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rybicki</surname>
          </string-name>
          . “
          <article-title>The great mystery of the (almost) invisible translator: Stylometry in translation”</article-title>
          . In: Studies in Corpus Linguistics. Ed. by
          <string-name>
            <given-names>M. P.</given-names>
            <surname>Oakes</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Ji</surname>
          </string-name>
          . Vol.
          <volume>51</volume>
          . Amsterdam: John Benjamins,
          <year>2012</year>
          , pp.
          <fpage>231</fpage>
          -
          <lpage>248</lpage>
          . doi:
          <volume>10</volume>
          .1075/scl.51.
          <year>09ryb</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rybicki</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Heydel</surname>
          </string-name>
          . “
          <article-title>The stylistics and stylometry of collaborative translation: Woolf's Night and Day in Polish”</article-title>
          .
          <source>InL:iterary and Linguistic Computing 28.4</source>
          (
          <issue>2013</issue>
          ), pp.
          <fpage>708</fpage>
          -
          <lpage>717</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqt02 7.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>C.</given-names>
            <surname>Schöch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dudar</surname>
          </string-name>
          , and E. Fileva, edsS.
          <source>urvey of Methods in Computational Literary Studies. Cls Infra</source>
          ,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .5281/zenodo.7892112.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>C.</given-names>
            <surname>Schöch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Patraș</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Santos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Erjavec</surname>
          </string-name>
          . “
          <article-title>Creating the European Literary Text Collection (ELTeC): Challenges and Perspectives”</article-title>
          .
          <source>InM: odern Languages Open</source>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .3828/mlo.v0i0.
          <fpage>364</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>P. W. H.</given-names>
            <surname>Smith</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Aldridge</surname>
          </string-name>
          . “
          <article-title>Improving Authorship Attribution: Optimizing Burrows' Delta Method”</article-title>
          .
          <source>In:Journal of Quantitative Linguistics 18.1</source>
          (
          <issue>2011</issue>
          ), pp.
          <fpage>63</fpage>
          -
          <lpage>88</lpage>
          . doi:
          <volume>10</volume>
          .1080/09296174.
          <year>2011</year>
          .
          <volume>533591</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [26]
          <string-name>
            <surname>E. Stamatatos. “</surname>
          </string-name>
          <article-title>A survey of modern authorship attribution methods”</article-title>
          .
          <source>IJno:urnal of the American Society for Information Science and Technology 60.3</source>
          (
          <issue>2009</issue>
          ), pp.
          <fpage>538</fpage>
          -
          <lpage>556</lpage>
          . doi:
          <volume>10</volume>
          .1002/asi.21001.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>