=Paper= {{Paper |id=Vol-3878/57_main_long |storemode=property |title=Hits or Misses? A Linguistically Explainable Formula for Fanfiction Success |pdfUrl=https://ceur-ws.org/Vol-3878/57_main_long.pdf |volume=Vol-3878 |authors=Giulio Leonardi,Dominique Brunato,Felice Dell'orletta |dblpUrl=https://dblp.org/rec/conf/clic-it/LeonardiBD24 }} ==Hits or Misses? A Linguistically Explainable Formula for Fanfiction Success== https://ceur-ws.org/Vol-3878/57_main_long.pdf
                                Hits or Misses? A Linguistically Explainable Formula for
                                Fanfiction Success
                                Giulio Leonardi1,*,† , Dominique Brunato2,† and Felice Dell’Orletta2,†
                                1
                                    University of Pisa
                                2
                                    Istituto di Linguistica Computazionale “Antonio Zampolli”, ItaliaNLP Lab, Pisa


                                                  Abstract
                                                  This study presents a computational analysis of Italian fanfiction, aiming to construct an interpretable model of successful
                                                  writing within this emerging literary domain. Leveraging explicit features that capture both linguistic style and semantic
                                                  content, we demonstrate the feasibility of automatically predicting successful writing in fanfiction and we identify a set of
                                                  robust linguistic predictors that maintain their predictive power across diverse topics and time periods, offering insights into
                                                  the universal aspects of engaging storytelling. This approach not only enhances our understanding of fanfiction as a genre
                                                  but also offers potential applications in broader literary analysis and content creation.

                                                  Keywords
                                                  fanfiction, Italian corpus, success prediction, linguistic features, Explainable Boosting Machine



                                1. Introduction and Motivation                                                                             machine learning offer a powerful lens for making ex-
                                                                                                                                           plicit patterns that may explain the complex interplay
                                The growing proliferation of online literary content has                                                   between reader engagement and content success.
                                led to the emergence of new genres and storytelling                                                           This paper moves in this field and presents a computa-
                                forms, with fanfiction being particularly popular among                                                    tional analysis focused on Italian fanfiction, addressing
                                teens and young adults. Fanfiction consists of stories                                                     the following research questions: i.) Can the success of
                                created by fans (mostly hobby authors) that extend or                                                      Italian fanfiction be automatically predicted using stylis-
                                alter the narrative of existing popular media like books,                                                  tic and lexical features of the texts?; ii.) Which types of
                                movies, comics or games, and represents a significant                                                      features demonstrate the highest predictive capability,
                                portion of user-generated content on the web [1]. In re-                                                   and how consistent are these features across different
                                cent years, the widespread popularity that this genre has                                                  time periods and thematic domains?; iii.) To what ex-
                                assumed has prompted research into the linguistic and                                                      tent can these features be explained in terms of their
                                stylistic elements that contribute to its success, mirror-                                                 contribution to predicting success?
                                ing studies conducted on more traditional literary genres                                                     Our contributions. i.) We collected a corpus of Ital-
                                [2, 3, 4], among others.                                                                                   ian fanfiction stories enriched with metadata considered
                                   Understanding the elements that contribute to narra-                                                    as proxies of their success; ii.) We investigate the relation-
                                tive success is a fascinating area of research with implica-                                               ship between stylistic and lexical features of stories and
                                tions across various fields, from literary analysis to digital                                             their success from a modeling perspective; iii.) We iden-
                                humanities. From a socio-linguistic perspective, it can                                                    tified the most influential features in success prediction,
                                offer deeper insights into people and culture. It also has                                                 showing the key role played by form and stylistic related
                                significant applications in areas such as personalized con-                                                features across time and thematic domains of fanfictions.
                                tent recommendation and educational technology [5, 6].                                                        The paper is structured as follows: Section 2 briefly
                                While personal interests undoubtedly play a crucial role                                                   contextualizes our study among relevant literature; Sec-
                                in predicting a reader’s engagement with a literary con-                                                   tion 3 presents the reference corpus of Italian fanfiction
                                tent, the way information is presented can also evoke                                                      stories that we collected; in Section 4 we provide an
                                different reactions and levels of interaction, ultimately                                                  overview of the approach we devised including the de-
                                influencing the narrative’s success. In this regards, recent                                               scription of features used for classification and the classi-
                                advancements in Natural Language Processing (NLP) and                                                      fiers employed. Section 5 discusses the main findings and
                                                                                                                                           offers a fine-grained analysis of the classification results
                                CLiC-it 2024: Tenth Italian Conference on Computational Linguistics,                                       in terms of feature explainability. In Section 6 we sum-
                                Dec 04 — 06, 2024, Pisa, Italy
                                *
                                  Corresponding author.
                                                                                                                                           marize key findings and outlining promising directions
                                †
                                  These authors contributed equally.
                                                                                                                                           for future research in this field.
                                $ g.leonardi5@studenti.unipi.it (G. Leonardi);
                                dominique.brunato@ilc.cnr.it (D. Brunato);
                                felice.dellorletta@ilc.cnr.it (F. Dell’Orletta)
                                            © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License
                                            Attribution 4.0 International (CC BY 4.0).




CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
2. Related Work                                             techniques is still limited. Mattei et al. [14] employ lin-
                                                            guistic profiling to analyze a corpus of Italian fanfiction
The exploration of online content and its engagement inspired by the Harry Potter series, with the purpose of
levels has increasingly benefited from advancements in identifying linguistic patterns associated with success.
NLP and machine learning. Different perspectives have Inspired by this previous study, our research aims to ex-
been touched upon considering different textual domains, tend these findings through a computational modeling
typology of linguistic features and quantitative metrics approach, investigating the power of linguistic features
to operationalize a very subjective concept like success. for predicting fanfiction success and their generalization
The study by Toubia and colleagues [7] explores how across different experimental settings.
the structure of narratives, particularly the internal se-
mantic progression measured by features derived from
dense word representations, affects the success of stories 3. Corpus Construction
across different text typologies (movies, TV shows, and
academic papers). Berger and colleagues [8] examine As a first step, we compiled a reference corpus of Ital-
how the linguistic structure of online content affects user ian fanfiction. To this end, we searched available texts
engagement, specifically by modeling sustainable atten- on efpfanfic.net, one of the largest Italian websites dedi-
tion. This concept goes beyond just attracting a reader cated to publishing and reading amateur stories, focusing
with a catchy headline or advertisement; it also encom- specifically on stories labeled in the fanfiction genre.
passes the likelihood that a reader will continue viewing      Using a web scraping system, we extracted fanfictions
or reading the content. In their analysis of more than      based  on the Harry Potter series, a highly popular fandom
35,000 online contents from heterogeneous sources, they     on  the site, boasting 57,196 stories published between
emphasize the role of features related to processing ease 2003 and 2023. Figure 1 presents the temporal distribution
and emotional language.                                     of these fanfictions up to 2020.
   In the realm of literary works, Ashok et al. [2] first      Additionally, we gathered a secondary corpus consist-
leverage stylometric analysis and machine learning tech- ing of 2,441 stories based on The Lord of the Rings series.
niques to predict the success of popular English novels This secondary corpus served as a test set to assess the
from the Gutenberg Project. Their approach demon- influence of thematic domains on the analysis of story
strated the potential of these techniques for assessing success.
literary success. Extending these findings, Maharajan          For this study, we focused on the first chapter of each
et al. [9] proposed a multi-task approach to simultane-     fanfiction  to ensure a consistent analysis. While it is
ously evaluating success and genre prediction. Using        widely  recognized   that thematic units within stories —
deep learning representations, in addition to hand-craft particularly the beginnings and endings — often differ
features related to topic, sentiment, writing style, and from the middle sections due to their distinct narrative
readability of books, they obtained better performance roles, we observed that the majority of stories (69%) con-
than the single success prediction task approach. Focus- sist of only a single chapter, making them effectively self-
ing on contemporary English-language literature, the contained. The efpfanfic portal allows users to review
study by Bizzoni and colleagues [10] investigate how per- each chapter with ratings marked as negative, neutral, or
ceived novel quality is influenced by a broad spectrum of positive. Consistent with prior research such as [9] we
textual features — such as those related to readability and used the absolute number of reviews to define the success
sentiment — and how these perceptions vary depending of a story, which we consider broadly as popularity. This
on the reader’s level of expertise.                         approach is based on the assumption that a high num-
   The growing volume of online fanfiction has also been ber of interactions, regardless of their sentiment, reflects
the subject of numerous studies, either from the perspec- strong reader’s engagement. This is especially confirmed
tive of text mining by using NLP or through a qualita- since in our dataset negative reviews represent less than
tive lens via a manual examination. A comprehensive 1% of the total.
survey of analyses in this direction has been recently         To formulate our success prediction task, we estab-
provided by [11]. For example, Milli and Bamman [12]        lished a review threshold to classify each story as either
explore the relationship between fanfiction and its orig- a success or a failure. After analyzing the distribution of
inal canon, offering one of the first empirical analyses reviews for Harry Potter texts (Figure 2), we decided to
of this genre. Similarly, Sourati et al. [13] find that the exclude stories that fell in the middle of the distribution –
similarity between fanfictions and their original stories those that could not be clearly defined as successes or fail-
— particularly in terms of emotional arcs and character ures. Consequently, stories with fewer than two reviews
dynamics—correlates significantly with fanfiction’s pop- (25th percentile) were classified as failures, and those
ularity.                                                    with more than six reviews (75th percentile) as successes.
   In the context of Italian fanfiction, research using NLP Stories within the interquartile range were excluded from
Table 1                                                           success grounded on interpretable factors, we decided
Descriptive Statistics for the Harry Potter (HP) and Lord of      to leverage explicit features modelling both style-related
The Rings (LOTR) Corpora                                          and lexical aspects of text as input for the classification
                                                                  system. To evaluate the effectiveness and robustness of
 Corpus      #texts    #negatives     #positives    avg. #tok
                                                                  these features, we conducted experiments across three
   HP        26,032      13,058         12,974        1911        conceptually distinct scenarios to evaluate the ability to
  LOTR        932         526            406          1946        discriminate success in different contexts. Specifically,
                                                                  the first scenario is in-domain: the classifier is evaluated
                                                                  on texts within the same thematic domain as the training
                                                                  set, using 10-fold cross-validation on the HP corpus. The
                                                                  second scenario is out-domain: the classifier is evalu-
                                                                  ated on texts from a different thematic domain than the
                                                                  training set. In this case, the HP corpus is used as the
                                                                  training set, while the LOTR corpus serves as the test set.
                                                                     Finally, in the cross-time scenario, the temporal im-
                                                                  pact on classification is considered. The classifier is
                                                                  trained solely on texts from the HP corpus published
                                                                  in 2011 and sequentially tested on texts from each other
                                                                  year from 2003 to 2020. The 2011 texts were chosen for
                                                                  training because this year has the largest amount of data
Figure 1: Distribution of all fanfictions from the Harry Potter   (3,755 texts), is approximately central within the tempo-
corpus by year of publication (up to 2020).
                                                                  ral range [2003, 2020], and is particularly significant for
                                                                  fanfiction production due to the release of the final film
                                                                  in the Harry Potter saga.
                                                                     The main components of our approach are detailed in
                                                                  the following sections.

                                                                  4.1. Success Predictors
                                                                  A comprehensive set of features was extracted for each
                                                                  story in the corpus. These features were categorized into
                                                                  two primary groups: linguistic features, reflecting the
                                                                  text’s linguistic style and structure and lexical features,
                                                                  representing the semantic content of the text.
Figure 2: Distribution of published fanfiction from the Harry
Potter corpus by number of reviews in the first chapter.          4.1.1. Linguistic Features
                                                            To model text’s linguistic style and structure, we drew in-
                                                            spiration from the linguistic profiling framework, a NLP-
the analysis. We also excluded texts published after 2020, based methodology in which a large set of linguistically-
considering them too recent for meaningful comparison. motivated features automatically extracted from anno-
   As summarized in Table 1, the final corpora, hereafter tated texts is used to obtain a vector-based representa-
abbreviated as HP (Harry Potter) and LOTR (The Lord of tion of it. Such representations can be then compared
the Rings), consist of 26,032 and 932 texts, respectively. across texts representative of different textual genres
                                                            and varieties to identify the peculiarities of each [15].
                                                            For our study, we relied on Profiling-UD1 , a multilin-
4. Methodology                                              gual tool inspired by this framework, which extracts over
                                                            130 linguistic features from texts using the Universal De-
Based on the newly collected dataset and its internal
                                                            pendencies (UD) annotation formalism. As described in
distinction, we formulated the task of success prediction
                                                            Brunato et al. [16], these features encompass a range of
as a binary classification problem, that is: given a story,
                                                            linguistic phenomena that can be classified into distinct
the model is asked to predict whether it belongs to the
                                                            groups covering e.g. shallow text features (e.g. document
successful or unsuccessful class, where the two classes
                                                            and sentence length, average word length), distribution
were defined according to the metric based on the number
                                                            of grammatical categories, inflectional morphology and
of reviews received by readers.
   In line with our main purpose to construct a model of 1 http://linguistic-profiling.italianlp.it/
syntactic properties related to local and global parse tree   Table 2
depth structure.                                              Classification Accuracy(%) of the Models. ‘Ling.’ and ‘Lex.’
   These features have proven effective in tasks related      refer respectively to models trained on linguistic and lexical
to modeling text form, such as assessing text complex-        features. The baseline corresponds to the majority class label.
ity, and identifying stylistic traits of authors or author         Scenario      SVM Ling. EBM Ling. SVM Lex. Baseline
groups. Building on previous research on a similar cor-
pus of fanfiction [14], we hypothesize that these features       in-domain          65.03      66.15       69.95     50.16
can also distinguish between successful and unsuccessful        out-domain          59.22      64.70       43.45     56.43
                                                               avg. cross-time      62.02      62.81       49.31     49.20
fanfictions from a modeling perspective.
                                                                   average          62.09      64.55       54.24     51.93

4.1.2. Lexical Features
The second representation employed is based on lexi-          better accuracy compared to linear models. Additionally,
cal information and leverages the relative frequency of       with a reasonable number of features, the model remains
n-grams in each document. The choice of n-grams, in           explainable. Each shape function can be visualized as
contrast to more powerful semantic representation de-         a two-dimensional plot, with the feature value on the
rived from embeddings, is deliberately motivated by the       x-axis and the score assigned by the shape function on
desire to use lexical features that remain completely ex-     the y-axis. A score greater than 0 indicates a contribution
plicit. The model, henceforth referred to as the Lexical      towards the positive class, whereas a score less than 0
Model, consists of the following features:                    indicates a contribution towards the negative class. The
                                                              final prediction value for a record is simply the sum of
     • Forms: unigrams, bigrams, and trigrams of to-          the scores obtained from each shape function, potentially
       kens.                                                  transformed by the link function. Beyond analyzing in-
     • Lemmas: unigrams, bigrams, and trigrams of lem-        dividual shape functions, the average contribution of
       mas.                                                   each feature can be evaluated by taking the mean of the
     • Characters: sequences of characters at the be-         absolute values of the assigned scores.
       ginning or end of words, ranging from 1 to 4              There are various algorithms within the family of
       characters in length.                                  GAMs, primarily distinguished by the method used to
                                                              fit the shape functions. In the case of the EBM, stan-
4.2. Classifiers                                              dard gradient boosting is used. However, in each boost-
                                                              ing iteration, the algorithm sequentially cycles through
In line with our research questions, the explainability       each feature, constructing each univariate shape function
of the classification is crucial to evaluate the impact of    through bagged boosted trees. This method has proven
linguistic and lexical features on the prediction of suc-     to be one of the most effective for training a GAM.
cess. Therefore, two classification algorithms that allow        For our study, the EBM was employed exclusively for
for a precise global explanation of the predictions were      experiments based on linguistic features due to the ex-
selected.                                                     cessive dimensionality of the lexical model. This high
   The first classifier employed is a linear Support Vector   dimensionality would have rendered the GAM too com-
Machine. By fitting a decision hyperplane in the feature      plex to interpret and too time-expensive to train.
space, this method enables the examination of the hy-
perplane’s coefficients to assess the importance of the
features.                                                     5. Results and Discussion
   The second algorithm employed is the Explainable
Boosting Machine (EBM), which belongs to the family of       The classification results are summarized in Table 2, for
Generalized Additive Models (GAMs). As explained in          each model and scenario under evaluation.
[17] a GAM is a model of the form:                              For models using linguistic features, in the in-domain
                                                             scenario both the SVM and the EBM outperform the ma-
                                                             jority class baseline, with accuracies of 65.03% and 66.15%
                               ∑︁
                 𝑔(𝑦) = 𝛽0 +        𝑓𝑛 (𝑥𝑛 )            (1)
                                                             respectively, compared to 50.16% for the baseline. This
   where 𝑔(.) is called the link function, used to model indicates that both classifiers are effectively capturing
the output (e.g., the logistic function for classification). the linguistic patterns associated with success within the
Each 𝑓𝑛 (.) is referred to as a shape function, which is a same thematic domain.
univariate function modeling the relationship between           For linguistic models, in the out-domain scenario the
the feature 𝑛 and the target.                                performance of the SVM drops significantly, with an ac-
   The prediction is thus a sum of 𝑛 non-linear and arbi- curacy of 59.22%, whereas the EBM experiences a less
trarily complex shape functions, generally resulting in
                                                                 tures. We provide an in-depth analysis of this model in
                                                                 the following section.

                                                                 5.1. The Model of Success
                                                                 To gain a better understanding of the classification results
                                                                 and identify the most influential features for predicting
                                                                 success, we ranked the features according to the absolute
                                                                 value of their weight in the EBM classifier model trained
                                                                 on the entire training set. Table 3 presents an extract of
Figure 3: Classification Accuracy in the Cross-Time Setting
                                                                 the top 15 features. The analysis reveals that, in addi-
                                                                 tion to basic text features such as the average document
                                                                 length (measured in tokens [1]) and the average word
drastic decline, achieving an accuracy of 64.70%. How-           length (in characters [2]), more complex linguistic prop-
ever, both classifiers still perform better than the baseline,   erties play a crucial role. Among these, features related
suggesting some degree of ability to generalize of the lin-      to verbal predicates and verbal morphology emerge as
guistic features across different thematic domains.              particularly influential. This suggests that the syntac-
   The lexical model, in the in-domain scenario, achieves        tic and morphological characteristics of verbs, such as
an accuracy of 69.56%, outperforming all models with lin-        tense, mood and person, provide valuable information
guistic features, suggesting that lexical features provide       for the classifier prediction, highlighting the importance
a more powerful representation for in-domain success             of deeper linguistic structures in building a model of
prediction. Nevertheless, in the out-domain scenario, the        successful writing.
lexical model does not surpass the baseline, indicating             While this ranking highlights the ‘global’ importance
a complete lack of predictive ability. This suggests that        of features, it does not explain their effect on classifica-
lexical features, which are primarily based on the content       tion. For a more detailed analysis, Figure 4 in Appendix
of the specific fanfiction’s narrative universe, perform         A highlights the threshold values for each of the top
well within the same thematic domain but lose all sig-           15 ranked features, indicating the point at which the
nificance outside of it. Conversely, linguistic features,        expected classification shifts from one class to another.
which focus on the form of the text, appear to be more           Additionally, it provides the number of instances in the
adaptable regardless of the theme.                               training set for each feature value. Interestingly, there
   Figure 3 presents the performance over time for classi-       are some features which split almost exactly the amount
fiers trained with linguistic features. Additionally, two        of data into two subsets. For example, the features rep-
baselines are shown: "Random Choice", which randomly             resenting word length (char_per_tok) has a discriminant
selects between the two classes, and "Maj. Class", which         threshold of 4.55 characters which distinguishes success-
always assigns the majority class from the correspond-           ful stories – typically with longer words – from unsuc-
ing training set (2011 stories), i.e. the positive one. The      cessful ones – usually with shorter words. Similarly, fea-
results of the lexical model in the cross-time scenario          tures related to the (morpho-)syntactic profile of the text
were insignificant, as they were very similar to the "Maj.       such as the percentage of conjunctions (dep_dist_conj)
Class" baseline. The classifier, therefore, defaults to as-      and non-finite verb forms (verbs_form_dist_Fin) show a
signing the negative class, demonstrating no predictive          similar pattern. For these features, values lower than the
capability. To avoid confusion, the lexical model results        discriminant threshold contribute to predicting the nega-
are not included in this Figure. In contrast, the cross-         tive class, effectively splitting the data into two groups
time results for models using linguistic features are more       with comparable densities. Regarding verb presence (ver-
meaningful: the results remain stable around an average          bal_head_per_sentence), an increased use of verbs corre-
of 62%, regardless of the dominant class in the tested           lates with the unsuccessful class. This finding contradicts
year and the classifier used (avg. cross-time in Table 2).       the idea that higher readability, typically conveyed by a
   The cross-time scenario further suggests that linguistic      predominantly verbal prose rather than a nominal one,
features possess greater adaptability beyond their own           is a good indicator of writing quality. However, it aligns
domain, maintaining a considerable degree of general-            with observations by Ashok et al. [2], who identified
ization over time. Conversely, lexical features seem func-       similar patterns in canonical literary novels.
tional only within the specific domain of the training set,         Features related to verbal morphology also show a
losing all predictive power for texts from different do-         peculiar trend. For instance, a complementary perspec-
mains. Overall the model that performed best on average          tive emerges concerning the use of person morphology.
across the three scenarios, and with the least variance          Increasing the use of second person plural beyond a rela-
in performance, is the EBM trained with linguistic fea-          tively low threshold (0.4) positively affects the prediction
of success, which may indicate an alignment with the impact of the author’s popularity and productivity on
Reader-Insert2 format, a specific type of fanfiction where the success of their fanfiction.
the reader assumes the role of the protagonist, heavily
relying on second-person narration. In contrast, an ex-
cessive use of the first person plural is associated with References
the negative class.
                                                            [1] K. Hellekson, K. Busse, Fan fiction and fan com-
                                                                 munities in the age of the internet: new essays,
Table 3                                                          McFarland, 2014.
Top 15 Scores of the EBM Trained with Linguistic Features   [2] V. G. Ashok, S. Feng, Y. Choi, Success with style:
         #     feature                         score             Using writing style to predict the success of novels,
                                                                 in: Proceedings of the 2013 conference on empirical
         #1    n_tokens                        0.121
                                                                 methods in natural language processing, 2013, pp.
         #2    char_per_tok                    0.098
         #3    verbal_root_perc                0.095
                                                                 1753–1764.
         #4    verbs_num_pers_dist_Plur+2 0.090             [3] J. Brottrager, A. Stahl, A. Arslan, U. Brandes,
         #5    verbs_num_pers_dist_Plur+1 0.088                  T. Weitin, Modeling and predicting literary recep-
         #6    upos_dist_SYM                   0.080             tion. a data-rich approach to literary historical re-
         #7    n_sentences                     0.077             ception, Journal of Computational Literary Studies
         #8    aux_tense_dist_Imp              0.077             1 (2022). URL: https://doi.org/10.48694/jcls.95.
         #9    verbs_tense_dist_Imp            0.072        [4] M. Algee-Hewitt, S. Allison, M. Gemma, R. Heuser,
        #10 aux_tense_dist_Pres                0.067             F. Moretti, H. Walser, Canon/archive : large-scale
        #11 verbal_head_per_sent               0.066             dynamics in the literary field, 2018. URL: https://
        #12 dep_dist_conj                      0.065
                                                                 litlab.stanford.edu/LiteraryLabPamphlet11.pdf.
        #13 tokens_per_sent                    0.064
                                                            [5] Reviews matter: How distributed mentoring
        #14 verbs_form_dist_Fin                0.053
        #15 n_prepositional_chains             0.052             predicts lexical diversity on fanfiction.net, 2018.
                                                                 URL: https://api.semanticscholar.org/CorpusID:
                                                                 265096028.
                                                            [6] S. Sauro, Fan fiction and informal language learning,
6. Conclusion                                                    The handbook of informal language learning (2019)
                                                                 139–151.
Understanding success factors in literary writing is an     [7] O. Toubia, J. A. Berger, J. Eliashberg, How quan-
evolving area of cross-disciplinary research. This study         tifying the shape of stories predicts their success,
on Italian fanfiction demonstrated the feasibility of pre-       Proceedings of the National Academy of Sciences of
dicting success using computational methods and ex-              the United States of America 118 (2021). URL: https:
plainability techniques. Notably, we found that features         //api.semanticscholar.org/CorpusID:235648521.
related to style and structure of texts show greater ro- [8] J. A. Berger, W. W. Moe, D. A. Schweidel, What
bustness than lexical ones across different domains and          holds attention? linguistic drivers of engagement,
time periods. This suggests that the way a story is crafted      Journal of Marketing 87 (2023) 793 – 809. URL: https:
may be more universally appealing than specific word             //api.semanticscholar.org/CorpusID:255250393.
choices or thematic elements.                               [9] S. Maharjan, J. Arevalo, M. Montes, F. A. González,
   We believe that the implications of this study extend         T. Solorio, A multi-task approach to predict lika-
far beyond fanfiction research. On the one hand, it pro-         bility of books, in: Proceedings of the 15th Confer-
vides new methodologies for analyzing online literary            ence of the European Chapter of the Association for
phenomena offering potential contributions to digital hu-        Computational Linguistics: Volume 1, Long Papers,
manities. From the NLP perspective, it could inform text         2017, pp. 1217–1227.
generation models, potentially guiding the creation of [10] Y. Bizzoni, P. F. Moreira, I. M. S. Lassen, M. R. Thom-
content that resonates more effectively with readers.            sen, K. Nielbo, A matter of perspective: Build-
   Future research could explore the generalizability of         ing a multi-perspective annotated dataset for the
these findings to other languages and genres, as well            study of literary quality, in: N. Calzolari, M.-Y.
as the investigation on the dynamics of evolving reader          Kan, V. Hoste, A. Lenci, S. Sakti, N. Xue (Eds.),
preferences over time by also considering alternative            Proceedings of the 2024 Joint International Con-
measures to gauge success. Additionally, this study does         ference on Computational Linguistics, Language
not take into account the importance of the author; a            Resources and Evaluation (LREC-COLING 2024),
potential future development would be to consider the            ELRA and ICCL, Torino, Italia, 2024, pp. 789–800.
2
                                                                 URL: https://aclanthology.org/2024.lrec-main.71.
https://fanlore.org/wiki/Reader-Insert
[11] D. Nguyen, S. Zigmond, S. Glassco, B. Tran, P. J.          A. Top 15 Features of the EBM
     Giabbanelli, Big data meets storytelling: using ma-
     chine learning to predict popular fanfiction, Social
     Network Analysis and Mining 14 (2024) 58.
[12] S. Milli, D. Bamman, Beyond canonical texts: A com-
     putational analysis of fanfiction, in: J. Su, K. Duh,
     X. Carreras (Eds.), Proceedings of the 2016 Confer-
     ence on Empirical Methods in Natural Language
     Processing, Association for Computational Lin-
     guistics, Austin, Texas, 2016, pp. 2048–2053. URL:
     https://aclanthology.org/D16-1218. doi:10.18653/
     v1/D16-1218.
[13] Z. Sourati Hassan Zadeh, N. Sabri, H. Chamani,
     B. Bahrak, Quantitative analysis of fanfictions’ pop-
     ularity, Social Network Analysis and Mining 12
     (2022) 42.
[14] A. Mattei, D. Brunato, F. Dell’Orletta, The style of
     a successful story: a computational study on the
     fanfiction genre, in: J. Monti, F. Dell’Orletta, F. Tam-
     burini (Eds.), Proceedings of the Seventh Italian
     Conference on Computational Linguistics, CLiC-it
     2020, Bologna, Italy, March 1-3, 2021, volume 2769
     of CEUR Workshop Proceedings, CEUR-WS.org, 2020.
     URL: https://ceur-ws.org/Vol-2769/paper_52.pdf.
[15] H. van Halteren, Linguistic profiling for authorship
     recognition and verification, in: Proceedings of the
     42nd Annual Meeting of the Association for Com-
     putational Linguistics (ACL-04), Barcelona, Spain,
     2004, pp. 199–206. URL: https://aclanthology.org/
     P04-1026. doi:10.3115/1218955.1218981.
[16] D. Brunato, A. Cimino, F. Dell’Orletta, G. Venturi,
     S. Montemagni, Profiling-UD: a tool for linguis-
     tic profiling of texts, in: N. Calzolari, F. Béchet,
     P. Blache, K. Choukri, C. Cieri, T. Declerck, S. Goggi,
     H. Isahara, B. Maegaard, J. Mariani, H. Mazo,
     A. Moreno, J. Odijk, S. Piperidis (Eds.), Proceedings
     of the Twelfth Language Resources and Evaluation
     Conference, European Language Resources Associ-
     ation, Marseille, France, 2020, pp. 7145–7151. URL:
     https://aclanthology.org/2020.lrec-1.883.
[17] Y. Lou, R. Caruana, J. Gehrke, Intelligible mod-
     els for classification and regression, Proceedings
     of the ACM SIGKDD International Conference
     on Knowledge Discovery and Data Mining (2012).
     doi:10.1145/2339530.2339556.
Figure 4: Visualization of the Shape Functions of the Top 15 Linguistic Features of the EBM. In each graph pair, the x-axis
represents the feature value, the y-axis of the line plot indicates the score assigned by the shape function, and the marked
threshold value denotes the feature value at the zero score point. For the features represented by absolute numbers (i.e.
n_tokens, char_per_tok, n_sentences, and n_prepositional_chains), the values are displayed as raw counts. For the remaining
features, which are expressed as percentage distributions, the values are shown accordingly. More details about how these
features are calculated are reported in [16].