<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Style Change Detection by Threshold Based and Window Merge Clustering Methods</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Neuch</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>tel</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The goal of the Style Change Detection task is to detect the stylistic changes in a document and exploit them to determine the number of authors. Our approach is to segment the given document into chunks of text, called windows. Each window is converted into feature vectors of words (around 50 features). The mutual distances among the window feature vectors are measured (using di erent distance measures like Matusita, Tanimoto etc.) and fed to two clustering algorithms called Threshold Based (TBC) and Window Merge (WMC). The number of clusters yielded correspond to the number of authors. The results of the two algorithms are then compared against each other and a combined minimum model.</p>
      </abstract>
      <kwd-group>
        <kwd>Authorship Multiple Authorship Style Change Detection Clustering</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Multiple authorship detection has long been a fascinating subject of research in
the literary community. Over time, the applications of multiple authorship have
extended into areas like plagiarism detection, forensics and more recently fake
news detection. It has been shown by multiple researchers ([
        <xref ref-type="bibr" rid="ref2">2</xref>
        ],[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]), how word
patterns can discern an exclusive personal style. Presence of multiple consistent
personal styles indicate presence of multiple authors.
      </p>
      <p>As part of the CLEF Style Change Detection problem, the model was tasked
to identify how many authors had written a given document or a stream of text.
More details can be found in [17].The dataset was composed of forum posts from
di erent `StackExchange' network sites. All documents were written in English.
The topics of discussion varied greatly. As the documents were based on forums,
each document was independent of others. Therefore, there were no separate
training sets for each author. In the previous years, viz. 2018 and 2017, the
style change detection problem at CLEF was limited to detecting the presence
or absence of style changes. It is to be noted that the Style Change Detection
is a variation of the Authorship Diarization problem introduced in 2016, which
aimed to cluster the contributions of each individual author of a document, when
the number of authors is either known or unknown.</p>
      <p>The rest of this paper is organized as follows. Section 2 describes the dataset.
In Section 3 we describe feature extraction, distance measurement and
clustering algorithm. Section 4 shows some of our evaluation results. A conclusion in
Section 5 draws the main ndings of our experiments.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Corpus</title>
      <p>In this section, we perform some preliminary statistics on the data to gain a
comprehensive understanding of the data. A basic statistical overview of the PAN
dataset is provided in Section 2.1 while the e ect of duplication is highlighted
in Section 2.2.
2.1</p>
      <sec id="sec-2-1">
        <title>Overall Statistics</title>
        <p>The training dataset consists of 2546 documents while the validation set consists
of 1272 documents. The number of authors ranges from 1 to 5. Each document
is composed of responses to a single thread at a forum. The mean number of
tokens was 1570 (median: 1452, sd: 810, min: 339, max: 7042). The punctuations
were preserved and treated as separate tokens.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Duplicate Sentences</title>
        <p>It was observed that certain documents contain duplicate sentences. In Figure 1,
we show a boxplot of duplicated sentences per document, grouped by the number
of authors. It can be noted that for documents with a single author, the number
of duplicates is rather low.</p>
        <p>On the other hand, the documents written by multiple authors have varying
number of duplicate sentences. In Table 1, we compare the number of
documents containing no duplicate sentences across di erent author classes. Since,
the number of duplicate sentences is low for single authors and relatively high
for multiple authors, therefore, the number of duplicate sentence may be used
as a feature in identifying single author documents.
This section details the implementation details of our method. We de ne our
feature selection methods, distance measures and clustering algorithm.
To preserve the logical document structure, we break down the document into
paragraphs. A window was initially assumed to be equal to a paragraph of text.
However, such a simple approach created highly variable lengths of windows.
Thus, the window tokenizer was modi ed to merge extremely short paragraphs
(less than 200 chars) with the previous paragraphs. Similarly, large paragraphs
were split into smaller windows to get reasonably length wise balanced windows.</p>
        <p>
          Thereafter, top 50 Most Frequent Tokens (MFT) were selected from each
document (and not the whole corpus). Each window was converted to a feature
vector based on the normalized frequency of the selected tokens. We used the
Matusita [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] distance measure to compare the distances between the window
feature vectors. Say, a document has 4 windows, then a symmetric 4x4 matrix
will be created representing the distance between the windows, as shown in
Table 2.
Our solution used a combination of two algorithms. We introduce a few de
nitions before introducing the algorithms.
        </p>
        <p>De nitions
{ Distance Matrix: A distance matrix represents the distances among all
the windows in a square matrix. The diagonal elements are zero because
they represent the distances between a window and itself. Table 2 shows a
distance matrix representing the relationships in Figure 3. For the purpose
of this paper, we will only deal with symmetric distance matrices.
{ Sorted Distance Array: Such an array contains non-duplicated entries
of the distance matrix, sorted in ascending order. In terms of a graph, an
entry in the distance matrix corresponds to the edge distance d between
nodes Wi and Wj . The top right half of the distance matrix (cells of Table 2
highlighted in gray) are extracted and converted to an array of the form (Wi,
Wj , d) where Wi, Wj are windows and d is the distance between them. An
example of sorted distance array is shown in Table 3.
{ Cluster: The most similar windows of text are grouped together into
clusters. The average distance of a cluster is the sum total of all the distances
between members divided by the number of edges. For example, in Figure 2,
the average distance of cluster 1 is 0.2/1 or 0.2.
{ Cluster List: Cluster list contains all the active clusters. Ideally, each cluster
corresponds to an author implying that the author has written the text in
the windows. Therefore, the number of clusters correspond with the number
of authors. However, in some cases, this is not true. For example, it may so
happen that an author has written only a few sentences which are contained
by a single window. Such a window, if su ciently distant from other clusters,
is likely to be excluded by all clusters. As a result, there is a possibility that
the number of authors calculated is lower than the ground truth. On the
other hand, if the document is broken down into tiny windows, it is possible
that there are too few characters, i.e not su cient information. As a result,
the distance measurement among such windows will be higher, resulting
in the formation of a higher number of small clusters. Thus the predicted
number of authors is higher than the ground truth.
{ Thresholds : The following thresholds are required by the Threshold Based
Clustering algorithm (TBC) to determine if a new member may be added
to an existing cluster or if two clusters may be merged together. Thresholds
ensure that existing clusters expand reasonably.</p>
        <p>Add Window Threshold: A new member may be added to a cluster,
if upon adding the new member, the revised average distance of the
modi ed cluster is not greater than x% of the average distance of the
existing cluster.</p>
        <p>Merge Cluster Threshold: Two clusters may be merged together, if
the revised average distance of the new cluster is not greater than x% of
the average distance of either of the existing clusters.</p>
        <p>However, xing a certain threshold as discussed above does not take into
account the size of the existing cluster. It may be desirable that, clusters
with only a few members or having a small average distance, expand more
rapidly as compared to larger clusters. Therefore, size of the cluster and its
average distance was used to penalise the respective threshold values of the
clusters. Such an adaptation, helped to customize the threshold according
to the state of the cluster.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Threshold Based Clustering Algorithm The intuition behind the TBC</title>
        <p>algorithm is that a good way to start clustering is to select the closest windows in
terms of distance because such windows, are likely to belong to the same cluster.
As such, the distance matrix is transformed into an array sorted by distance, viz.
the sorted distance array or dist arr and fed to the Algorithm 1. The idea is to
select the windows with smallest distances iteratively such that, when forming
a cluster, the closest members are included rst. Thereafter, the members that
Algorithm 1 Threshold Based Clustering Algorithm
1: procedure CreateCluster(dist arr) # Convert sorted distance array to
clusters
2: cluster list
3: for all entry e 2 dist arr do # where e is of the form (Wi; Wj ; d)
4: Wi; Wj GetWindows(e)
5: if Wi 2= c and Wj 2= c0 8 c; c0 2 cluster list then # where c is a cluster
6: c fWi; Wj g
7: cluster list cluster list [ c
8: end if
9: if 9 c0 : c0 2 cluster list and Wi 2 c0 and Wj 2= c 8c 2 cluster list then
10: c00 c0 [ fWj g
11: if average distance(c00) &lt; add node threshold then
12: c0 c00 # cluster c' is updated by adding a new Window Wj
13: end if
14: end if
15: if 9 c0 : c0 2 cluster list and Wj 2 c0 and Wi 2= c 8c 2 cluster list then
16: c00 c0 [ fWig
17: if average distance(c00) &lt; add window threshold then
18: c0 c00 # cluster c' is updated by adding a new Window Wi
19: end if
20: end if
21: if 9 c; c0 : c; c0 2 cluster list and Wi 2 c and Wj 2 c0 then
22: c00 c [ c0
23: if average distance(c00) &lt; merge cluster threshold then
24: cluster list cluster list [ fc00g fc; c0g
25: # Replace previous smaller clusters with a new merged cluster
26: end if
27: end if
28: end for
29: return len(cluster list) # Number of clusters represent number of
authors
30: end procedure
are further away, are included. Hence the clusters would be expected to grow
from small and dense to large and spread out. Such a growth is controlled by
the thresholds shown in Section 3.2. For each entry in the dist arr, the windows
may have any one of the following conditions:
{ None of the windows are present in any existing cluster of the cluster list.</p>
        <p>Therefore, a new cluster must be created with these two windows and added
to the cluster list.
{ One of the windows is already a member of an certain existing cluster, while
the other isn't. Thus, the non-member window may be added to an existing
cluster, provided that the Add Window Threshold condition is met. While
adding a new member, all the corresponding edges between the new member
and existing members are extracted from the dist arr and added to the
cluster.
{ Both windows belong to two di erent existing clusters which may be merged
if the conditions of the Merge Cluster Threshold are satis ed. As in the
previous case, all the corresponding edges are added to the cluster
It is to be noted, that the condition that both windows belong to the same
cluster is not possible, as otherwise, such an entry would already have been
traversed while adding new members or cluster merging.
3.3</p>
      </sec>
      <sec id="sec-2-4">
        <title>Window Merge Clustering</title>
        <p>The Window Merge Clustering algorithm (WMC) iteratively combines the most
similar windows to generate a new set of windows. From these new windows, the
distance matrix is re-calculated for the next iteration. As a result, the clusters
formed are hierarchical. The idea is to represent each cluster with a combined
representation of all of its members together, rather than individual distances.
Figure 4 shows the formation of the clusters by the Window Merge Clustering.
The height of the cluster corresponds to the distance between the two clusters.
In the rst iteration, Cluster1 (W0, W1) is formed and W0 and W1 are
combined. The resulting new set of windows are W0W1, W2 and W3. In the next
iteration Cluster2 (W2,W3) is formed. The updated windows are W0W1 and
W2W3. Eventually, these windows are connected in the third iteration resulting
in Cluster3 (W0W1, W2W3).</p>
        <p>To determine the number of clusters, we use a truncation threshold, which is
a certain percentage of the height of the largest cluster. In this example shown
in Figure 4, the truncation threshold is set to 0.43 or 75% of 0.57 which is the
height of Cluster3. With the help of the truncation threshold, we can control
the size of the clusters. Also, as in the case of the TBC, the number of clusters
relates to the number of authors.</p>
        <p>Algorithm 2 Window Merge Algorithm
1: procedure CreateHierarchicalCluster(dist matrix; text windows)
2: window merge order
3: while size(dist matrix) &gt; 1 do
4: Wi; Wj FindMinDistanceWindows(dist matrix)
5: merge order window merge order [ fWi; Wjg
6: new text windows MergeWindows(Wi; Wj; text windows)
7: dist matrix CalculateDistMatrix(new text windows)
8: end while
9: authors Truncate(window merge order; truncation threshold)
10: return authors
11: end procedure
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <p>The Add Window and Merge Cluster thresholds of the TBC algorithm were
varied from 5% to 100%. We observed the optimum performance at 50% for
both the thresholds. In our experiments, both thresholds are set to 50%. We
show the evaluation of our algorithms in Table 4. In addition to TBC and WMC,
we have another model called the combined minimum. The combined minimum
model simply selects the smaller predicted value from either of TBC and WMC
algorithms. In our evaluation, we nd that the the TBC performs better than
the WMC and the combined minimum.</p>
      <p>In Section 2.2, we proposed the idea of using the number of duplicates
sentences to predict the number of authors. We showed how low number of duplicate
sentences were indicative of a single author. We modi ed the algorithms TBC
and WMC such that, if no duplicate sentences were present in a document, only
a single author was predicted. However, if one or more duplicate sentences were
observed, the algorithm was executed normally. In Table 5, the impact of using
such a feature is shown. We observe that the performances of all the three
models have improved signi cantly. In Table 5, the results of the test set evaluated
on the TIRA for TBC algorithm are shown.
In this paper, we have demonstrated that it is possible to identify the number
of authors from a relatively short piece of text without any training corpus per
author. We have contributed two di erent algorithms to perform authorship
clustering and now we present the chief ndings of our work.</p>
      <p>We found that, TBC perfoms the best out of the three models under both
the scenarios. This is because by prioritizing the selection of the smallest
distances, we ensured that the most vital members of a cluster are selected rst.
Such members are vital because they lay down the cluster structure. Subsequent
members can only be included if they expand the cluster within a reasonable
limit. We observe that WMC performs only slightly worse, when no information
on duplicate sentences is present. This shows that there is scope of improvement
in the combined representation approach of WMC.</p>
      <p>Also, by including the duplicate sentences feature, the performances of the
algorithms were improved by a considerable margin. Since, we have tested
using only one test collection, there may be some unknown dataset characteristic
favouring one algorithm over another. As part of our future work, one or two
additional datasets would be included for a better evaluation.</p>
      <p>Another important part of the window based approaches is to tune the
correct size of the window. Ideally the document should be split such that each
window contains exactly one style change completely. However, is not possible
to know the location of the style changes initially. Hence it is di cult to know
the correct size of a window. A possible improvement of our work could be to
de ne a measure for cluster quality achieved to be given as a feedback for further
improving the splitting of the windows in an iterative fashion.</p>
      <p>
        The github repository for our approach can be found at [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
16. Sebastiani, F.: Machine learning in automated text categorization. ACM
computing surveys (CSUR) 34(1), 1{47 (2002)
17. Zangerle, E., Tschuggnall, M., Specht, G., Potthast, M., Stein, B.: Overview of the
Style Change Detection Task at PAN 2019. In: Cappellato, L., Ferro, N., Losada,
D., Muller, H. (eds.) CLEF 2019 Labs and Workshops, Notebook Papers.
CEURWS.org (Sep 2019)
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Burrows</surname>
          </string-name>
          , J.:
          <article-title>All the way through: testing for authorship in di erent frequency strata</article-title>
          .
          <source>Literary and Linguistic Computing</source>
          <volume>22</volume>
          (
          <issue>1</issue>
          ),
          <volume>27</volume>
          {
          <fpage>47</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Craig</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kinney</surname>
            ,
            <given-names>A.F.</given-names>
          </string-name>
          :
          <article-title>Shakespeare, computers, and the mystery of authorship</article-title>
          . Cambridge University Press (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Crystal</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Language and the internet</article-title>
          . Cambridge University Press (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Daelemans</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kestemont</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manjavancas</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Specht</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tschuggnall</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zangerle</surname>
          </string-name>
          , E.: Overview of PAN 2019:
          <article-title>Author Pro ling, Celebrity Pro ling, Cross-domain Authorship Attribution and Style Change Detection</article-title>
          . In: Crestani,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Braschler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Savoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Rauber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            , Muller, H.,
            <surname>Losada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Heinatz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Cappellato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Ferro</surname>
          </string-name>
          , N. (eds.)
          <source>Proceedings of the Tenth International Conference of the CLEF Association (CLEF</source>
          <year>2019</year>
          ). Springer (Sep
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Harman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>How e ective is su xing? Journal of the american society for information science 42(1</article-title>
          ),
          <volume>7</volume>
          {
          <fpage>15</fpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. Je reys, H.:
          <article-title>Theory of Probability,</article-title>
          .
          <source>OUP Oxford</source>
          (
          <year>1948</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kocher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Savoy</surname>
          </string-name>
          , J.:
          <article-title>Distance measures in author pro ling</article-title>
          .
          <source>Information Processing &amp; Management</source>
          <volume>53</volume>
          (
          <issue>5</issue>
          ),
          <volume>1103</volume>
          {
          <fpage>1119</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Nath</surname>
            ,
            <given-names>S.: SCD CLEF</given-names>
          </string-name>
          <year>2019</year>
          . https://github.com/pan-webis-de/nath19 (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Pennebaker</surname>
            ,
            <given-names>J.W.:</given-names>
          </string-name>
          <article-title>The secret life of pronouns: What our words say about us</article-title>
          . Bloomsbury Press, New York (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gollub</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>TIRA Integrated Research Architecture</article-title>
          . In: Ferro,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <surname>C</surname>
          </string-name>
          . (eds.)
          <article-title>Information Retrieval Evaluation in a Changing World - Lessons Learned from 20 Years of</article-title>
          CLEF. Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>A decade of shared tasks in digital text forensics at pan</article-title>
          .
          <source>In: European Conference on Information Retrieval</source>
          . pp.
          <volume>291</volume>
          {
          <fpage>300</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montes-y Gomez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Overview of the 6th author pro ling task at pan 2018: multimodal gender identi cation in twitter</article-title>
          .
          <source>Working Notes Papers of the CLEF</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Savoy</surname>
          </string-name>
          , J.:
          <article-title>Comparative evaluation of term selection functions for authorship attribution</article-title>
          .
          <source>Digital Scholarship in the Humanities</source>
          <volume>30</volume>
          (
          <issue>2</issue>
          ),
          <volume>246</volume>
          {
          <fpage>261</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Savoy</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Analysis of the style and the rhetoric of the 2016 us presidential primaries</article-title>
          .
          <source>Digital Scholarship in the Humanities</source>
          <volume>33</volume>
          (
          <issue>1</issue>
          ),
          <volume>143</volume>
          {
          <fpage>159</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Savoy</surname>
          </string-name>
          , J.:
          <article-title>Is starnone really the author behind ferrante? Digital Scholarship in the Humanities 33(4</article-title>
          ),
          <volume>902</volume>
          {
          <fpage>918</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>