<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Models of Literary Evaluation and Web 2.0. An Annotation Experiment with Goodreads Reviews</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Simone</forename><surname>Rebora</surname></persName>
							<email>simone.rebora@univr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Verona</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gabriele</forename><surname>Vezzani</surname></persName>
							<email>gabriele.vezzani@univr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Verona</orgName>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">RWTH Aachen University</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Models of Literary Evaluation and Web 2.0. An Annotation Experiment with Goodreads Reviews</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">069C9BCF83B252228EBEAB6131831DA4</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:47+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>literary evaluation</term>
					<term>digital social reading</term>
					<term>annotation</term>
					<term>transformer models</term>
					<term>LLMs</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the context of the Web 2.0, user-genrated reviews are becoming more and more prominent. The particular case of book reviews, often shared through digital social reading platforms such as Goodreads or Wattpad, is of particular interest, in that it offers scholars data regarding literary reception of unprecedented size and diversity. In this paper, we test whether the evaluative criteria employed in Goodreads reviews can be included in the framework of traditional literary criticism, by combining literary theory and computational methods. Our model, based on the work of von Heydebrand and Winko, is first tested through the practice of heuristic annotation. The generated dataset is then used to train a Tranformer-based classifier. Last, we compare the performance of the latter with that obtained by instructing a Large Language Model, namely GPT-4.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Nowadays, reviews are ubiquitous. Pushed by our natural tendency to share information and encouraged by the very companies that sell us their products, we constantly take part in the production and accumulation of huge amounts of data regarding our preferences and judgements. Literature has also been strongly affected by this phenomenon. On digital social reading platforms <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b18">19]</ref>, such as Goodreads, one can find terabytes of information regarding the reception of millions of books, covering the tastes of the most diverse typologies of readers. Finding efÏcient ways to manage these records is essential not only for market reasons -as demonstrated by the great effort that companies put into developing algorithms to better capture and predict users' tastes 1 -but also for research purposes. Never before, in fact, has it been possible to gather insights about the reception of literary works as extensive and diverse as the ones that can be drawn from these platforms. Designing ways to analyze this data will allow us to shed a whole new light on the dynamics that regulate literary evaluation.</p><p>Literary studies have been preoccupied with evaluation since around the 1980s, when, thanks to the influence of the emergent field of postcolonial studies <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b24">25]</ref>, it started to become more and more evident that the prominence of the works of the so-called western canon wasn't at all objective, but rather an arbitrary construct, contingent on the social and cultural background of the readers. Van Rees <ref type="bibr" target="#b27">[28]</ref> proposed a model in which literary value was built through a three-step process involving journalistic reviewers, essayists, and academic critics. At the same time more detailed and flexible, von Heydebrand and Winko's model <ref type="bibr" target="#b12">[13]</ref> also posits the social construction of literary value, although allowing for the contribution of any agent within the literary field. Similar ideas are shared by Dalen-Oskam <ref type="bibr" target="#b5">[6]</ref>, who recently demonstrated, in an empirical fashion, how social biases can influence the evaluation of books.</p><p>According to these authors, society influences literary evaluation by setting up standards against which books are judged. This would imply the existence of criteria independent from any specific evaluative act, that manifest themselves across a multiplicity of such acts, thus offering a means to systematize them. In fact, as von Hydebrand and Winko have shown, although they can vary greatly depending on the historical context, these "standards of value" can nonetheless be organized according to some general features: they can concern the aesthetic level of a book (its style, characters, plot, etc.), the way it relates to other works (thus being original or derivative, for instance), the impact it has on its reader, and so on.</p><p>In this paper, we build on von Heydebrand and Winko's theory and interpret online reviews as acts of "linguistic evaluation". What distinguishes such acts from more implicit forms of evaluation (even the simple purchase of a book could be interpreted as such) is that they "require a standard of value -as well as certain categorizing assumptions -in order to progress from the description of a text to its evaluation" <ref type="bibr">[12, p. 227</ref>]. Seen as evaluative acts of this kind, book reviews are not simple expressions of individual taste, but the result of a socially learned practice with specific schemes and regularities. Different kinds of evaluative criteria can then be taken as the axis along which to situate single reviews, thus quantifying their variability and reducing their complexity. Furthermore, knowing which criteria are implemented by an individual (or a group) could allow us to predict their future judgements or to reveal the societal influences operating on them.</p><p>To make the identification of evaluative criteria as objective as possible, we turned to the practice of "interpretative markup", a form of annotation "devoted to recording a scholar's or analyst's observations and conjectures in an open-ended way" <ref type="bibr">[20, p. 202</ref>]. As argued by Gius and Jake <ref type="bibr" target="#b9">[10]</ref>, this kind of annotation, when carried out in a collaborative way, can be a powerful tool for systematizing the interpretation of potentially ambiguous texts. Our main goal consisted in using the annotated dataset for training a classifier capable of recognizing the use of different evaluative criteria in possibly any online book review.</p><p>The paper is organized as follows. In section 2, we present the corpus that constituted the basis of our work. The discussion of the annotation process occupies Section 3, where we outline the tagset we developed (3.1), the workflow we followed and the main problems we encountered <ref type="bibr">(3.2)</ref>. To automate the annotation process we tried two different approaches: finetuning a Bert-based classifier and instructing a Large Language Model (GPT 4). The results of these approaches are presented, respectively, in Sections 4.1 and 4.2. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Data</head><p>The dataset used to perform the annotation task was built on the same basis of the AbsORB dataset: a selection of "approximately six million English language reviews of nine different genres (i.e., fantasy, romance, thriller, horror, mystery, science fiction, historical fiction, contemporary, classics)" <ref type="bibr" target="#b14">[15]</ref>, downloaded from the Goodreads website between 2018 and 2019. Out of them, a total of 100 reviews were randomly selected, with just one filter set to exclude extremely short (below 200 words) ones. To counter an intrinsic imbalance in the dataset (i.e., the dominance of reviews of "young adult" novels, frequently tagged as "fantasy" or "romance"), we added a filter which forced the selection of 20 reviews from among the ones tagged as "classics". Reviews were then automatically split into sentences by using SpaCy<ref type="foot" target="#foot_0">2</ref> , so as to allow a sentence-by-sentence annotation.</p><p>A first partitioning of our corpus, composed by 11 reviews, was used as a toy dataset to train the annotators, and was not therefore retained for analysis. The remaining 89 reviews, which constituted our main corpus, had a mean length of 1155 tokens (SD = 449). They were written by 84 different reviewers in a timespan going from 2008 to 2018. The number of reviews per year is lightly skewed towards more recent dates reflecting the growth of the Goodreads platform. There are 84 different books reviewed in the corpus, spanning across 9 different genres (with most of the books belonging to more than one genre). The distribution of reviews per genre and per year can be seen, respectively, in Figures <ref type="figure" target="#fig_1">2 and 1</ref>.</p><p>All copyright and privacy implications in using such a dataset have already been discussed <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b22">23]</ref>. In any case, to safeguard the rights of the authors of the reviews and to comply to copyright limitations (while still profiting from the research exceptions recently introduced in multiple European legislations, following the 2019 Directive on Copyright in the Digital Single Market), we decided not to share it publicly. Researchers who would like to access it will have to contact the authors of this paper, by stating their intended use. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Annotation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Designing the Tag Set</head><p>The starting point for the creation of our tag set was von Heydebrand and Winko's model, which systematizes literary evaluation based on the different kinds of criteria that can be employed for such a task. However, in order to be adapted to our material and goals, it needed to be simplified. First of all, as demonstrated by the high number of categories they devised, it is clear that von Heydebrand and Winko aim at maximizing the exhaustivity of their model. Our intent was slightly different. In fact, we needed to account for the highest possible number of scenarios with the fewest possible categories, in order to reach a number of examples per category high enough to train a classifier, without having to annotate an extreme number of reviews.</p><p>Furthermore, despite their intent to take into consideration the contributions of potentially any actor within the literary field, it is quite clear that the kind of evaluative acts that von Heydebrand and Winko had in mind while designing their model were quite different from what one can find today on digital social reading platforms (and quite understandably so, given that their essay predates the advent of the latter by almost ten years). Many of their categories reflect the judgment patterns of, if not professionals, at least cultivated readers, capable of assessing the salient stylistic features of a given book, as well as its positioning in the overall landscape of literary tradition. On the other hand, online reviews "are mostly expressions of consumer satisfaction or dissatisfaction" <ref type="bibr">[26, p. 3]</ref>. In them, the individual dimension is magnified, with evaluations hinging around the reader's personal relation with the book.</p><p>In our tag set, we devoted a special attention to criteria based on one's own thoughts, feelings and personal experiences. However, not all can be traced back to the individual level. By contrast, we believe literary evaluation to be a complex phenomenon, involving not only a reader and a book, but also the whole community in which the interaction between the former two takes place. To account for such complexity, we have incorporated in our tag set two further dimensions alongside that of individual criteria: one for evaluations based on the book itself, its content-related or formal aspects, and one for considerations on its societal value, its impact on a given community of readers. Furthermore, we have included a label for evaluations that do not fall into any of the aforementioned categories. In Appendix A, we list the 7 labels of our tag set, with a short description for each. Figure <ref type="figure" target="#fig_2">3</ref>, on the other hand, shows the overall structure of the tag set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">The Annotation Workflow</head><p>After a brief training regarding the task they were going to perform, two annotators were asked to independently annotate all the reviews in our corpus, attributing one (and only one) label to each of their sentences. The choice to take sentences as the basic unit of our analysis allowed us to organize the annotation in a very straightforward way, by providing annotators with tables where each row corresponded to one sentence, to which they had to associate a label in a dedicated column (columns containing the title of the reviewed book, the review and the sentence id numbers were also included). In the cases where an evaluation spanned across multiple sentences, annotators simply had to attribute the same label to each one of them.</p><p>A first partitioning of our corpus, composed by 11 reviews, was used as a toy dataset to train the annotators (annotations were therefore not retained for analysis). The actual annotation work was carried out on the remaining 89 reviews in 4 consecutive rounds, each followed by a meeting during which cases of disagreement were discussed, as well as possible shortcomings of our tagging system. Inter-annotator agreement was computed using Cohen's Kappa. The scores for each round are reported in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Considering the entirety of the tag set ('all_labels'), the coefÏcients are in the range of a moderate agreement <ref type="bibr" target="#b17">[18]</ref>, while their increase over time can be taken as a sign of the efÏcacy of our workflow. Furthermore, the growth of the agreement for the tags 'generic_val' and 'aesthetic' can be seen as a consequence of the consecutive clarifications that we were able to offer during the meetings that followed each round. The agreement for the category 'no_val' (i.e., the simple distinction between evaluative and non-evaluative sentences), despite a drop in round 2, remained almost unchanged throughout the entirety of the work. It should be noted that this was by far the most frequent label, accounting for 26% of all annotations. This is why we managed to reach such a substantial coefÏcient for this category, which still registered many instances of disagreement, resulting for the most part from intrinsic ambiguities in our data. Take, for instance, even a sentence as simple as the following: "[this book] is not a horror novel" <ref type="foot" target="#foot_1">3</ref> . In and of itself, this would look like a mere observation. However, with a small interpretative leap, we could see it as a negative evaluation, a way for the reader to express their disappointment in finding out that the book was not what they expected it to be. To make the identification of evaluative sentences as unambiguous as possible, we then decided that for a sentence to be considered as such it needed to contain an explicit evaluation of the reviewed book, or an explicit mention of the impact it had on the reviewer. Although focusing only on explicit cases could be seen as limiting, such approach allowed us: a) to reduce disagreement between annotators, and b) to gather a corpus of clearly evaluative sentences to successfully train a classifier, if not in recognizing all of our categories, at least in distinguishing evaluative statements from other elements that can be found in a review, like summaries of the book's plot, accounts of reviewer's personal experiences, and so on.</p><p>The absence of agreement registered for the 'social' label in the first three rounds can be explained by the extremely limited number of occurrences of this category, which made it difÏcult to develop specific guidelines. To understand such scarcity, let's remember that the label was thought to account for references of socially established evaluative standards, such as prestigious prizes and canonical works. In many cases, such standards are the manifestation of an academic -as in the case of canonical works, see <ref type="bibr" target="#b10">[11]</ref> -, or professional perspective on literature. They are the standards held by that cohesive community of readers that Chervel calls the "literary establishment" <ref type="bibr" target="#b4">[5]</ref>. By not referencing the standards of the latter, Goodreads users simply demonstrate to belong to a different community, characterized by its own standards and rituals (think about the Goodreads Choice Award, a literary prize where winners are voted exclusively by members of the Goodreads community).</p><p>More than giving rise to a stable and universal canon, on Goodreads the social component of literary evaluation seems to operate at the level of specific literary genres. Indeed, both annotators reported noticing that reviews tended to follow genre-specific evaluative patterns. One of these patterns, concerning the over-represented genre of young adult novels (in Figure <ref type="figure" target="#fig_1">2</ref> split between romance and fantasy) could be at the heart of the problems we encountered with the labels in the 'individual' category, reflected in the erratic trends of the corresponding kappa co-efÏcients. Something that became apparent already from the first rounds of annotation, is that readers of this genre tend to develop very strong personal relationships with the characters in the books they read, either treating them as if they were real persons or completely identifying with them. This phenomenon gave rise to an ambiguity that caused a good part of the disagreement between annotators. Take, for instance, the sentence "there is a depth to his character that bit by bit when revealed you can't help but fall that much harder for him because his character demands nothing less" <ref type="foot" target="#foot_2">4</ref> . Apart from the poor grammar, one notes right away that the reviewer is evaluating the way a character is built, which would make the sentence fall under the category 'aesthetic'. However, far from being a neutral assessment of a stylistic feature, the sentence is highly emotionally charged, which would justify the label 'ind_emotional'. Lastly, based on the patent involvement of the reader in what they are reading, one could argue for the attribution of this sentence to the 'ind_pragmatic' category. To face such ambiguity, we decided to tag as 'aesthetic' all the sentences that contained an explicit reference to literary art (e.g., plot, characters, style).</p><p>The last step after annotation has been a curation phase, during which we 'settled' the cases of disagreement between the annotators and attributed a definitive label to the respective sentences. This was necessary to allow for the subsequent phase, the training of a classifier, for which each sentence in our dataset needed to be assigned to one and only one label. In extremely rare cases (less than one percent of the sentences) we also intervened to correct attributions that, despite the annotators' agreement, were blatantly wrong.<ref type="foot" target="#foot_3">5</ref>  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Traininig the Classifier. BERT vs. GPT</head><p>The following step in our project relied on the annotated dataset to train a classifier able to recognize evaluative acts in large amounts of text. In recent years, two main approaches have become dominant in machine learning for text classification. The first implies fine-tuning base (or task-specific) Transformer models thanks to the knowledge stored in annotated datasets <ref type="bibr" target="#b13">[14]</ref>. The second benefits from the flexibility of Large Language Models, instructing them to perform ex novo the annotation task based on a set of instructions (or examples) <ref type="bibr" target="#b20">[21]</ref>. The materials created in our project offered a valid groundwork for both approaches.</p><p>Analyses have been carried out with a series of Python scripts, which can all be consulted in the project's GitHub repository. <ref type="foot" target="#foot_4">6</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Fine-tuning Bert models</head><p>For the first approach, the 6,014 annotated and curated sentences were used as ground truth to fine-tune multiple Transformer models of the Bert family. We decided to test the following three models:</p><p>• google-bert/bert-base-uncased (from now on, referred to as google-bert), as representative of a large, general-purpose model for English language; • LiYuan/amazon-review-sentiment-analysis (from now on, LiYuan), because it was finetuned on a similar kind of data (general product reviews) for a similar kind of task (sentiment analysis); • JoelVIU/bert-base-uncased-finetuned-amazon_reviews_books (from now on, JoelVIU ), because it was finetuned on an even more coherent dataset (Amazon book reviews).</p><p>Given the already-discussed issues of unbalance and underrepresentation in the annotation labels, we decided to simplify them by adopting two different strategies:</p><p>our tag set and guidelines were addressed during the course of the annotation itself, thanks to the insights we gathered at each new step. 1. we reduced the labels to three classes: 'eval_individual' (under which the three categories related to the impact on the individual reviewer were merged), 'eval_generic' (under which all the evaluation categories were merged), and 'no_val'; 2. we further reduced the labels to two classes: 'eval' (under which all the evaluation/impact labels were merged) and 'no_val'.</p><p>Testing was first performed via a 5-fold cross validation, to establish which one among the chosen models was the most performant. As shown by Table <ref type="table" target="#tab_1">2</ref>, performance depended on the task, with the JoelVIU model slightly outperforming the others for the three-label classification and google-bert producing the best results for the binary classification. However, it should be noted that the three-label setup produced low F1-macro scores because of the substantial failure in classifying the 'eval_individual' sentences (F-1 scores for that label were never higher than 0.144 throughout models/folds), due probably to the very low number of available samples in the dataset (444 sentences, corresponding to 7.4% of the total). In light of these considerations, we decided to use the google-bert model as a reference point for our subsequent analyses.</p><p>A second level of analysis concerned the fitness of the number of annotated sentences to effectively train the classifier. To address this question, we performed a series of 5-fold cross validations by fine-tuning the model with an increasing number of sentences (from 600 to 6,000, selected randomly from the dataset). Figure <ref type="figure" target="#fig_3">4</ref> shows how, for both the binary and three-class classification, F1-macro scores reach a plateau at around 3,000 sentences, thus suggesting that the learning threshold for the classifier is substantially lower than the number of annotated sentences.</p><p>A third and final level of analysis was chosen in order to get efÏciency scores comparable to the ones obtained with Large Language Models (see section 4.2). Here the selection strategy was changed slightly, by randomly choosing entire reviews instead of single sentences, without performing any k-fold cross validation. After testing different configurations, we selected one with 18 reviews in the test set (corresponding precisely to 20% of the sentences in the dataset, with comparable distributions of labels), which produced efÏciency scores similar to the ones obtained before (for an overview, see Table <ref type="table" target="#tab_3">3 and 4</ref>).</p><p>Given the unsatisfactory performance reached on the three-labels dataset (the model simply ignored the under-represented 'eval_individual' category) we decided to share only the model finetuned for the binary classification of evaluative and non-evaluative sentences, which can now be freely accessed on Hugging Face.<ref type="foot" target="#foot_5">7</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Instructing GPT 4</head><p>For the second approach, the annotation guidelines were used as a starting point to develop system prompts for the GPT-4 Large Language Model. By situating itself in the emergent area of "prompt engineering" <ref type="bibr" target="#b3">[4]</ref>, the work here became more exploratory, aiming at the identification of the best technique to instruct the model.</p><p>In the first phase of testing, the overall prompting strategy was a simple zero-shot (i.e., instructions plus input), with system prompts (i.e., the instructions) composed by adopting three different approaches:</p><p>• The first, defined as complex, was a straightforward adaptation of the annotation guidelines in their latest stage. Note that guidelines were loosely structured, with occasional repetitions and multiple addenda; • The second, defined as simple, was a drastic simplification of the annotation guidelines, extracting only the most relevant information and giving it a simpler structure; • The third, defined as procedural, was inspired by the chart in Figure <ref type="figure" target="#fig_2">3</ref> and by the "Tree of Thoughts" (ToT) prompting technique <ref type="bibr" target="#b3">[4]</ref>, producing a more structured prompt, where the assignment of a label was the result of a set of nested choices.</p><p>These three approaches were then combined with three different tagsets:</p><p>• The first, defined as full, adopted the 7 labels originally used by the annotators (note that this setup was not even tested with Bert models, because of the scarcity of annotations for some labels); • The second, defined as 3-class, reduced the labels to the first setup with Bert models: 'eval_individual', 'eval_generic', and 'no_val'; • The third, defined as binary, further reduced the labels to the second Bert setup: 'val' and 'no_val'.</p><p>For each approach, the system prompt produced with the full tagset was then adapted into 3-class and binary by performing the minimum amount possible of modifications, so as to keep the core of the prompt intact. Final result was a set of 9 different system prompts, which can be consulted in Appendix B.</p><p>To profit from the ability of GPT-4 to process large amounts of text and to emulate as closely as possible the work of the annotators, we decided to give it as an input entire reviews split into sentences. The user prompt was therefore structured as a .csv file with three columns: 'book_title' (containing the title of the reviewed book), 'sentence_id' (with a numeric identifier of the sentence), and 'sentence' (with the sentence text). At each trial, GPT-4 was prompted 18 times, with the 18 reviews identified with the criteria described in section 4.1.</p><p>Requests were processed by using the "gpt-4o-2024-05-13" model (with temperature set to 0 to produce the most deterministic behavior) on the OpenAI API. The whole operation cost a total of 3$. Comparisons of the GPT-4 annotations with the ground truth are shown by Table <ref type="table" target="#tab_4">5</ref>. Note that the adaptation of the full tagset to 3-class and binary, and of the 3-class tagset to binary could also be accomplished ex post (i.e., after having performed the annotation with GPT-4).</p><p>Overall, GPT-4 efÏciency is higher than fine-tuned Bert models for the 3-class condition (in fact, GPT-4 performs substantially better on the 'eval_individual' label, overcoming the issue of its underrepresentation in the dataset), while it is lower for the binary condition. When comparing the different system prompts, the complex approach performs slightly better than the procedural one, suggesting how a structured prompt may not be necessary to obtain the highest efÏciency. Finally, ex-post tagset adaptation produces mixed results (with the best efÏciency for the 3-class and binary tagsets obtained with and without adaptation, respectively), even if in the majority of the cases (6 out of 9) it worsens efÏciency. Tables <ref type="table" target="#tab_6">6, 7</ref>, and 8 show more in detail the efÏciency of the best-performing setups.</p><p>In the second phase of testing, we adopted a "few-shot" strategy by using as a basis the best performing setup (complex approach with binary tagset). The "few-shot" strategy implies providing not only instructions, but also examples for performing the task. Examples were extracted from the remaining 71 reviews, selecting the ones that showed the proportion of 'val' vs. 'no_val' labels closest to the overall mean in the dataset (i.e. the most "balanced" ones). These reviews were presented to the model together with the curated annotations, before asking it to annotate the new reviews. Three different tests were then performed, with an increasing number of sample reviews:</p><p>• two reviews, corresponding to 172 sentences (the whole operation cost 0.66$);</p><p>• four reviews, corresponding to 354 sentences (1.21$);</p><p>• eight reviews, corresponding to 596 sentences (1.95$).</p><p>As prices increased substantially, we decided not to test bigger selections of sample reviews. Also, quite surprisingly, this prompting technique had a detrimental effect on the efÏciency of the model, with F1-macro scores decreasing to 0.780 (with two sample reviews), 0.754 (four reviews), and 0.724 (eight reviews).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>We believe that our work can contribute to the field in several ways. First, the development of a tag set to capture evaluative criteria in unstructured reviews enriches the current search for possible ways to operationalize literary evaluation and study it from an empirical perspective <ref type="bibr" target="#b26">[27,</ref><ref type="bibr" target="#b5">6]</ref>. Furthermore, the application of the aforementioned tag set during the work of annotation revealed some interesting features of the material analyzed. First, let's note that many scholars seem to interpret online book reviewing as animated by a distrust for (if not a full-fledged opposition to) more traditional practices of literary criticism, as embodied by bourdeauian "gatekeepers" <ref type="bibr" target="#b2">[3]</ref> such as publishers, professors or jurors of a literary prize. For Franzen, online critics are characterized by "a general distrust of established institutions of aesthetic opinion formation" <ref type="bibr">[9, p. 3]</ref>. Even when it is not so clearly thematized, a similar view of today's critical landscape is implicit in many works that aim at unveiling the differences between lay criticism and its traditional counterpart <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b28">29]</ref>. Last, scholars have interpreted the decline of journalistic criticism as an effect of the rise of internet's rating culture, which, by empowering readers and allowing them to express their own judgements, brings to the slow and inevitable demise of those figures once charged with directing the public's taste <ref type="bibr" target="#b4">[5]</ref>. Contrasting such interpretations, the reviewers in our corpus showed no concern for established critical discourse, as demonstrated by the extremely low number of instances of the 'social' label.</p><p>The detection of genre-specific evaluative patterns is another interesting finding that deserves to be further investigated in future works. Two interpretations for such result can be hypothesized: first, that the different features of books of different genres <ref type="bibr" target="#b7">[8]</ref> somehow 'call for' different evaluative approaches, or, second, that different literary genres have different social constituencies, that is, they are read by different 'kinds of persons' <ref type="bibr" target="#b16">[17]</ref>, who, in turn, elaborate different evaluative patterns. Unfortunately, our data do not allow us to effectively test either one of these hypotheses.</p><p>Last, many researchers are searching for computational ways to analyze the evaluation of books through user-generated online reviews <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b6">7]</ref>. We contribute to this particular line of research by developing two models for the classification, in online book reviews, of: a) evaluative and non-evaluative sentences and b) generic evaluations, evaluations based on the impact of the book on the reader, and non-evaluative sentences. Furthermore, the comparison between the performance of these models and that of and GPT-4 confirm the findings of recent studies <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b23">24]</ref>, demonstrating the validity of employing LLMs for classification tasks, where they can attain results that are comparable to those of fine-tuned Transformer models. Further research is needed on the reasons for the unsuccess of the "few-shot" prompting technique, which could equally be ascribed to limitations in LLMs or to the complexity of the annotation task. However, this result highlights the importance of applying LLMs to datasets like ours, which could pose new challenges and stimuli for their development.</p><p>• no_val: The sentence does not express an evaluation of the reviewed book • generic_val: Into this category fall all those evaluations related to the work as a whole ("beautiful book, " "highly recommended, " "an unbearable read, " and so on). Also, tag with this category all evaluative sentences that do not fall into any of the following categories. • aesthetic: Any evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...). • social: This value concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book).</p><p>Of particular interest here are all those judgments that seek to enact (or to reafÏrm) a canonization of the judged work, that is, to place it on the roster of 'important' readings. • ind_cognitive: Evaluations regarding the cognitive impact of a book on a reader, the information that the latter extracted from the former or the intellectual stimulation they experienced while reading it. • ind_pragmatic: Evaluations regarding the impact of the book on the reader's life or the existential "lessons" that the latter learned from the former • ind_emotional: Evaluations regarding the emotional impact of the book on the reader, the way the former made the latter feel.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. The System Prompts</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.1. complex_full</head><p>You will receive as input a .csv table with the following structure: book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>When the sentence does not express an evaluation of the reviewed book, assign the label "no_val".</p><p>When the sentence expresses an evaluation of the reviewed book, you will have to choose between six different labels.</p><p>1. "aesthetic": Any evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...).</p><p>The following three labels do not refer to features present in the text (such as formal values), but rather to the impact it had on the reader. They are divided into:</p><p>2. "ind_pragmatic": Evaluations regarding the impact of the book on the reader's life, the existential "lessons" that the latter learned from the former.</p><p>3. "ind_emotional": Evaluation regarding the emotional impact of the book on the reader, the way the former made the latter feel.</p><p>4. "ind_cognitive": Evaluation regarding the cognitive impact of a book on a reader, the information that the latter extracted from the former or the intellectual stimulation they experienced while reading it.</p><p>The last two labels are: 5. "social": This value concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that seek to enact (or to reafÏrm) a canonization of the judged work, that is, to place it on the roster of 'important' readings.</p><p>6. "generic_val": Into this category fall all those evaluations related to the work as a whole ("beautiful book, " "highly recommended, " "an unbearable read, " and so on). Also, tag with this category all evaluative sentences that do not fall into any of the above categories.</p><p>In assigning one (and only one) of these labels to each sentence, please follow these generic guidelines:</p><p>-Work on individual sentences. ONLY in cases where one sentence is incomprehensible without the next, or expresses a concept that necessarily requires continuation in the next, treat the two as a single block and assign them the same label.</p><p>-If judgments related to several categories are made in a sentence, tag it with the category that seems the most important to you.</p><p>-Judgements regarding other books than the one reviewed, or judgments related to past readings of the same book must be tagged as "no_val".</p><p>-What does NOT constitute an evaluation: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries.-"ind_pragmatic" must contain an explicit reference to the reviewer's real life or experience.</p><p>-"aesthetic" must have explicit reference to literary art (plot, characters, style).</p><p>-An evaluation must specify, explicitly, its object.</p><p>-An evaluation may have neutral, or ambiguous ('mixed feelings') valence.</p><p>-Any comparison, resulting in the priority of one over the other, between the book in question and other cultural products is to be considered evaluative.</p><p>-All references to story, characters, style, or any other features that relate back to the writing are tagged as aesthetic regardless of the simplicity of the rating. Example: I like the story, I like the characters, etc.</p><p>-"generic_val" consists of all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.2. complex_3-class</head><p>You will receive as input a .csv table with the following structure:</p><p>book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>When the sentence does not express an evaluation of the reviewed book, assign the label "no_val".</p><p>When the sentence expresses an evaluation of the reviewed book, you will have to choose between two different labels.</p><p>Assign the label "eval_generic" for any evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...).</p><p>The other label, "eval_individual" does not refer to features present in the text (such as formal values), but rather to the impact it had on the reader. It can be used for:</p><p>1. Evaluations regarding the impact of the book on the reader's life, the existential "lessons" that the latter learned from the former.</p><p>2. Evaluation regarding the emotional impact of the book on the reader, the way the former made the latter feel.</p><p>3. Evaluation regarding the cognitive impact of a book on a reader, the information that the latter extracted from the former or the intellectual stimulation they experienced while reading it.</p><p>The label "eval_generic" can also be interpreted in two additional ways:</p><p>1. This label concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that seek to enact (or to reafÏrm) a canonization of the judged work, that is, to place it on the roster of 'important' readings.</p><p>2. Into this category fall all those evaluations related to the work as a whole ("beautiful book, " "highly recommended, " "an unbearable read, " and so on). Also, tag with this category all evaluative sentences that do not fall into any of the above cases.</p><p>In assigning one (and only one) of these labels to each sentence, please follow these generic guidelines:</p><p>-Work on individual sentences. ONLY in cases where one sentence is incomprehensible without the next, or expresses a concept that necessarily requires continuation in the next, treat the two as a single block and assign them the same label.</p><p>-If judgments related to several categories are made in a sentence, tag it with the category that seems the most important to you.</p><p>-Judgements regarding other books than the one reviewed, or judgments related to past readings of the same book must be tagged as "no_val".</p><p>-What does NOT constitute an evaluation: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries.</p><p>-"eval_individual" must contain an explicit reference to the reviewer's real life or experience.</p><p>-An evaluation must specify, explicitly, its object.</p><p>-An evaluation may have neutral, or ambiguous ('mixed feelings') valence.</p><p>-Any comparison, resulting in the priority of one over the other, between the book in question and other cultural products is to be considered evaluative.</p><p>-All references to story, characters, style, or any other features that relate back to the writing are tagged as "eval_generic" regardless of the simplicity of the rating. Example: I like the story, I like the characters, etc.</p><p>-"eval_generic" can consist of all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.3. complex_binary</head><p>You will receive as input a .csv table with the following structure:</p><p>book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>When the sentence does not express an evaluation of the reviewed book, assign the label "no_val".</p><p>When the sentence expresses an evaluation of the reviewed book, you will have to use the label "val".</p><p>Assign this label for any evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...).</p><p>Assign the label "val" also when the evaluation does not refer to features present in the text (such as formal values), but rather to the impact it had on the reader. It can be used for:</p><p>1. Evaluations regarding the impact of the book on the reader's life, the existential "lessons" that the latter learned from the former.</p><p>2. Evaluation regarding the emotional impact of the book on the reader, the way the former made the latter feel.</p><p>3. Evaluation regarding the cognitive impact of a book on a reader, the information that the latter extracted from the former or the intellectual stimulation they experienced while reading it.</p><p>The label "val" can also be interpreted in two additional ways: 1. This label concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that seek to enact (or to reafÏrm) a canonization of the judged work, that is, to place it on the roster of 'important' readings.</p><p>2. Into this category fall all those evaluations related to the work as a whole ("beautiful book, " "highly recommended, " "an unbearable read, " and so on). Also, tag with this category all evaluative sentences that do not fall into any of the above cases.</p><p>In assigning one (and only one) of these labels to each sentence, please follow these generic guidelines:</p><p>-Work on individual sentences. ONLY in cases where one sentence is incomprehensible without the next, or expresses a concept that necessarily requires continuation in the next, treat the two as a single block and assign them the same label.</p><p>-If judgments related to several categories are made in a sentence, tag it with the category that seems the most important to you.</p><p>-Judgements regarding other books than the one reviewed, or judgments related to past readings of the same book must be tagged as "no_val".</p><p>-What does NOT constitute an evaluation: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries.</p><p>-An evaluation must specify, explicitly, its object.</p><p>-An evaluation may have neutral, or ambiguous ('mixed feelings') valence.</p><p>-Any comparison, resulting in the priority of one over the other, between the book in question and other cultural products is to be considered evaluative.</p><p>-All references to story, characters, style, or any other features that relate back to the writing are tagged as "val" regardless of the simplicity of the rating. Example: I like the story, I like the characters, etc.</p><p>-"val" can consist of all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.4. simple_full</head><p>You will receive as input a .csv table with the following structure:</p><p>book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines:</p><p>-"no_val", when the sentence does not evaluate the reviewed book; -"aesthetic", any evaluation concerning the specifics of literary language, both in its formal and content aspects;</p><p>-"ind_pragmatic", evaluations regarding the impact of the book on the reader's life; -"ind_emotional": it is about the value the reader places on the work based on what it made him or her feel. It can range from aspects more related to the book itself, to more intimate and personal issues;</p><p>-"ind_cognitive": in this category fall all considerations of a book's ability to teach the reader something or stimulate him intellectually;</p><p>-"social": this value concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book); -"generic_val": tag with this label all evaluative sentences that do not fall into any of the above categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.5. simple_3-class</head><p>You will receive as input a .csv table with the following structure: book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines:</p><p>-"no_val", when the sentence does not evaluate the reviewed book; -"eval_individual", evaluations regarding the impact of the book on the reader's life; the value the reader places on the work based on what it made him or her feel (it can range from aspects more related to the book itself, to more intimate and personal issues); all considerations of a book's ability to teach the reader something or stimulate him intellectually;</p><p>-"eval_generic", any evaluation concerning the specifics of literary language, both in its formal and content aspects; it also concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book); tag with this label all evaluative sentences that do not fall into any of the above categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.6. simple_binary</head><p>You will receive as input a .csv table with the following structure:</p><p>book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines:</p><p>-"no_val", when the sentence does not evaluate the reviewed book; -"val", evaluations regarding the impact of the book on the reader's life; the value the reader places on the work based on what it made him or her feel (it can range from aspects more related to the book itself, to more intimate and personal issues); all considerations of a book's ability to teach the reader something or stimulate him intellectually; any evaluation concerning the specifics of literary language, both in its formal and content aspects; it also concerns the impact that a book has had not on a single reader, but on a community of readers (references to literary awards, to the popularity of the book); tag with this label all evaluative sentences that do not fall into any of the above categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.7. procedural_full</head><p>You will receive as input a .csv table with the following structure: book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>Guideline 1: You will have to treat each sentence as a single unit of meaning. Therefore, you will assign a label based on the sentence alone. Only in cases where one sentence is incomprehensible without the next, treat the two as a single block and assign them the same label.</p><p>Guideline 2: You will have to assign one (and only one) label to each sentence. If several labels can be assigned to one sentence, choose the label that fits best with the sentence.</p><p>Guideline 3: Possible labels are: "no_val", "aesthetic", "ind_pragmatic", "ind_emotional", "ind_cognitive", "social", "generic_val". Labels can be grouped into two main categories: nonevaluative sentences and evaluative sentences. When assigning a label to a sentence, you will have to (1) identify the best-fitting category and (2) choose the best-fitting label.</p><p>Possible labels are categorized and described here below. Category 1: Non-evaluative sentences.</p><p>A non-evaluative sentence is a sentence that does not express an evaluation of the reviewed book, or that does not explicitly describe the impact it had on the reader.</p><p>Category 1, Label 1: "no_val" Use this label for all non-evaluative sentences. Use it also when the sentence expresses evaluations regarding other books than the one reviewed, or evaluations related to past readings of the reviewed book. Use it also in the case of: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries.</p><p>Category 2: Evaluative sentences. An evaluative sentence expresses an evaluation of the reviewed book, or it explicitly describes the impact it had on the reader. An evaluative sentence may have positive, negative, or ambiguous ('mixed feelings') valence. An evaluative sentence must specify, explicitly, its object. Any comparison, resulting in the priority of one over the other, between the reviewed book and other cultural products is to be considered as an evaluative sentence.</p><p>Category 2, Label 1: "aesthetic" Use this label when the sentence explicitly expresses an evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...). The evaluation must have explicit reference to literary art (e.g., plot, characters, style).</p><p>Category 2, Label 2: "ind_pragmatic" Use this label when the sentence explicitly describes the impact the reviewed book had on the reader. In particular, use it for sentences regarding the impact of the book on the reader's life, the existential 'lessons' that the reader learned from the book. The sentence must contain an explicit reference to the reviewer's real life or experience.</p><p>Category 2, Label 3: "ind_emotional" Use this label when the sentence explicitly describes the impact the reviewed book had on the reader. In particular, use it for sentences regarding the emotional impact of the book on the reader, the way the former made the latter feel. Still, be careful in not assigning this label to sentences that use an emotional language in a generic and/or metaphorical way (e.g. "I loved this book, " "I enjoyed the reading").</p><p>Category 2, Label 4: "ind_cognitive" Use this label when the sentence explicitly describes the impact the reviewed book had on the reader. In particular, use it for sentences regarding the importance of the information that the reader extracted from the book, or the intellectual stimulation the reader experienced while reading it. Category 2, Label 5: "social" Use this label when the sentence evaluates a book not based on the experience of a single reader, but on the experience of a community of readers (e.g., references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that enact (or reafÏrm) a canonization of the judged work, placing it on the roster of 'important' readings.</p><p>Category 2, Label 6: "generic_val" Use this label for all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader (e.g., "beautiful book, " "highly recommended, " "an unbearable read"). Also, tag with this category all sentences that do not fall into any of the above categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.8. procedural_3-class</head><p>You will receive as input a .csv table with the following structure:</p><p>book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>Guideline 1: You will have to treat each sentence as a single unit of meaning. Therefore, you will assign a label based on the sentence alone. Only in cases where one sentence is incomprehensible without the next, treat the two as a single block and assign them the same label.</p><p>Guideline 2: You will have to assign one (and only one) label to each sentence. If several labels can be assigned to one sentence, choose the label that fits best with the sentence. Guideline 3: Possible labels are: "no_val", "eval_generic", and "eval_individual". They are described here below. Label 1: "no_val" Use this label for all non-evaluative sentences. A non-evaluative sentence is a sentence that does not explicitly express an evaluation of the reviewed book, or that does not explicitly describe the impact it had on the reader. Use it also when the sentence expresses evaluations regarding other books than the one reviewed, or evaluations related to past readings of the reviewed book. Use it also in the case of: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries. Label 2: "eval_individual" Use this label when the sentence explicitly describes the impact the reviewed book had on the reader.</p><p>In particular, use it for sentences regarding the impact of the book on the reader's life, the existential 'lessons' that the reader learned from the book. The sentence must contain an explicit reference to the reviewer's real life or experience.</p><p>Use it also for sentences regarding the emotional impact of the book on the reader, the way the former made the latter feel. Still, be careful in not assigning this label to sentences that use an emotional language in a generic and/or metaphorical way (e.g. "I loved this book, " "I enjoyed the reading").</p><p>Finally, use it for sentences regarding the importance of the information that the reader extracted from the book, or the intellectual stimulation the reader experienced while reading it.</p><p>Label 3: "eval_generic" Use this label for all evaluative sentences. An evaluative sentence is a sentence that explicitly expresses an evaluation of the reviewed book. An evaluative sentence may have positive, negative, or ambiguous ('mixed feelings') valence. An evaluative sentence must specify, explicitly, its object. Any comparison, resulting in the priority of one over the other, between the reviewed book and other cultural products is to be considered as an evaluative sentence.</p><p>Use this label when the sentence explicitly expresses an evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...). The evaluation must have explicit reference to literary art (e.g., plot, characters, style).</p><p>In addition, use this label when the sentence evaluates a book not based on the experience of a single reader, but on the experience of a community of readers (e.g., references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that enact (or reafÏrm) a canonization of the judged work, placing it on the roster of 'important' readings.</p><p>Finally, use this label for all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader (e.g., "beautiful book, " "highly recommended, " "an unbearable read"). Also, tag with this category all evaluative sentences that do not fall into any of the above cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.9. procedural_binary</head><p>You will receive as input a .csv table with the following structure: book_title,sent_id,sentence The table includes the review of one book (identified by "book_title"), split into sentences. You will have to produce as output another .csv table with the following structure: sent_id,label You will assign the label to each sentence (even when the sentence is broken or incomplete) by following these guidelines.</p><p>Guideline 1: You will have to treat each sentence as a single unit of meaning. Therefore, you will assign a label based on the sentence alone. Only in cases where one sentence is incomprehensible without the next, treat the two as a single block and assign them the same label.</p><p>Guideline 2: You will have to assign one (and only one) label to each sentence. If several labels can be assigned to one sentence, choose the label that fits best with the sentence.</p><p>Guideline 3: Possible labels are: "no_val" and "val". They are described here below. Label 1: "no_val" Use this label for all non-evaluative sentences. A non-evaluative sentence is a sentence that does not explicitly express an evaluation of the reviewed book, or that does not explicitly describe the impact it had on the reader. Use it also when the sentence expresses evaluations regarding other books than the one reviewed, or evaluations related to past readings of the reviewed book. Use it also in the case of: interpretations ("the author wishes to express... " and the like), personal anecdotes ("I first read the book when I was in college"), plot summaries. Label 2: "val" Use this label for sentences expressing the impact of the reviewed book on the reader and for all evaluative sentences.</p><p>Use this label when the sentence explicitly describes the impact the reviewed book had on the reader. In particular, use it for sentences regarding the impact of the book on the reader's life, the existential 'lessons' that the reader learned from the book. The sentence must contain an explicit reference to the reviewer's real life or experience.</p><p>Use it also for sentences regarding the emotional impact of the book on the reader, the way the former made the latter feel. Still, be careful in not assigning this label to sentences that use an emotional language in a generic and/or metaphorical way (e.g. "I loved this book, " "I enjoyed the reading").</p><p>Finally, use it for sentences regarding the importance of the information that the reader extracted from the book, or the intellectual stimulation the reader experienced while reading it.</p><p>Use this label also for evaluative sentences. An evaluative sentence is a sentence that explicitly expresses an evaluation of the reviewed book. An evaluative sentence may have positive, negative, or ambiguous ('mixed feelings') valence. An evaluative sentence must specify, explicitly, its object. Any comparison, resulting in the priority of one over the other, between the reviewed book and other cultural products is to be considered as an evaluative sentence.</p><p>Use this label when the sentence explicitly expresses an evaluation concerning the specifics of literary language, both in its formal aspects (use of rhetorical figures, writing style, etc...) and content aspects (character or plot construction, narratological features, etc...). The evaluation must have explicit reference to literary art (e.g., plot, characters, style).</p><p>In addition, use this label when the sentence evaluates a book not based on the experience of a single reader, but on the experience of a community of readers (e.g., references to literary awards, to the popularity of the book). Of particular interest here are all those judgments that enact (or reafÏrm) a canonization of the judged work, placing it on the roster of 'important' readings.</p><p>Finally, use this label for all those statements not accompanied by explanation, i.e., expressions of appreciation not related to specific aspects of the book or to specific effects it had on the reader (e.g., "beautiful book, " "highly recommended, " "an unbearable read"). Also, tag with this category all evaluative sentences that do not fall into any of the above cases.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Distributions of books per genre</figDesc><graphic coords="3,193.46,84.17,208.36,153.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Distribution of reviews per year</figDesc><graphic coords="4,193.46,84.17,208.36,153.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Chart of the annotation workflow</figDesc><graphic coords="5,89.28,84.18,416.72,396.01" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: F1-macro scores for the google-bert model, depending on the number of available sentences.</figDesc><graphic coords="9,89.28,84.17,416.72,312.54" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Kappa values for different labels throughout the rounds</figDesc><table><row><cell>Label</cell><cell cols="4">Second round Third round Fourth round Fifth round</cell></row><row><cell>all_labels</cell><cell>0.49</cell><cell>0.46</cell><cell>0.51</cell><cell>0.57</cell></row><row><cell>no_val</cell><cell>0.66</cell><cell>0.57</cell><cell>0.64</cell><cell>0.64</cell></row><row><cell>aesthetic</cell><cell>0.43</cell><cell>0.52</cell><cell>0.49</cell><cell>0.52</cell></row><row><cell>ind_cognitive</cell><cell>0.25</cell><cell>0.18</cell><cell>0.16</cell><cell>0</cell></row><row><cell>ind_emotional</cell><cell>0.39</cell><cell>0.05</cell><cell>0.27</cell><cell>0.57</cell></row><row><cell>ind_pragmatic</cell><cell>0.25</cell><cell>0.17</cell><cell>0</cell><cell>0</cell></row><row><cell>social</cell><cell>0</cell><cell>0</cell><cell>0</cell><cell>0.33</cell></row><row><cell>generic_val</cell><cell>0.35</cell><cell>0.32</cell><cell>0.44</cell><cell>0.47</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Mean efficiency scores for the three models. A full report on the 5-fold cross validation can be found in our GitHub repository</figDesc><table><row><cell>Model google-bert</cell><cell cols="3">Metric Accuracy F1-macro F1-weighted 0.795 3-class Binary 0.820 0.860 0.553 0.832 0.860</cell></row><row><cell></cell><cell>Accuracy</cell><cell>0.809</cell><cell>0.848</cell></row><row><cell>LiYuan</cell><cell>F1-macro</cell><cell>0.528</cell><cell>0.818</cell></row><row><cell></cell><cell cols="2">F1-weighted 0.782</cell><cell>0.850</cell></row><row><cell>JoelVIU</cell><cell cols="2">Accuracy F1-macro F1-weighted 0.796 0.819 0.558</cell><cell>0.858 0.826 0.858</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Efficiency report for two classes on the sample dataset</figDesc><table><row><cell>Class</cell><cell cols="4">Precision Recall F1-score Support</cell></row><row><cell>no_val</cell><cell>0.895</cell><cell>0.933</cell><cell>0.913</cell><cell>854</cell></row><row><cell>val</cell><cell>0.817</cell><cell>0.731</cell><cell>0.772</cell><cell>349</cell></row><row><cell>Accuracy Macro avg Weighted avg</cell><cell>0.856 0.872</cell><cell>0.832 0.874</cell><cell>0.874 0.843 0.872</cell><cell>1203 1203 1203</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Efficiency report for three classes on the sample dataset</figDesc><table><row><cell>Class</cell><cell cols="4">Precision Recall F1-score Support</cell></row><row><cell>eval_generic</cell><cell>0.628</cell><cell>0.705</cell><cell>0.664</cell><cell>244</cell></row><row><cell>eval_individual</cell><cell>0.000</cell><cell>0.000</cell><cell>0.000</cell><cell>105</cell></row><row><cell>no_val</cell><cell>0.871</cell><cell>0.947</cell><cell>0.907</cell><cell>854</cell></row><row><cell>Accuracy Macro avg Weighted avg</cell><cell>0.500 0.746</cell><cell>0.551 0.815</cell><cell>0.815 0.524 0.779</cell><cell>1203 1203 1203</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>F1-macro scores for the different system prompts</figDesc><table><row><cell></cell><cell></cell><cell>full</cell><cell></cell><cell cols="2">3-class</cell><cell>binary</cell></row><row><cell>adapted ex post to</cell><cell>-</cell><cell cols="2">3-class binary</cell><cell>-</cell><cell>binary</cell><cell>-</cell></row><row><cell>complex simple procedural</cell><cell>0.425 0.336 0.386</cell><cell>0.634 0.531 0.675</cell><cell>0.758 0.717 0.797</cell><cell>0.668 0.630 0.641</cell><cell>0.803 0.801 0.807</cell><cell>0.825 0.779 0.803</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>Efficiency report for the best-performing prompt (complex_full) on the full tagset</figDesc><table><row><cell>Class</cell><cell cols="4">Precision Recall F1-score Support</cell></row><row><cell>aesthetic</cell><cell>0.312</cell><cell>0.863</cell><cell>0.458</cell><cell>131</cell></row><row><cell>generic_val</cell><cell>0.593</cell><cell>0.444</cell><cell>0.508</cell><cell>108</cell></row><row><cell>ind_cognitive</cell><cell>0.263</cell><cell>0.333</cell><cell>0.294</cell><cell>15</cell></row><row><cell>ind_emotional</cell><cell>0.365</cell><cell>0.613</cell><cell>0.458</cell><cell>75</cell></row><row><cell>ind_pragmatic</cell><cell>0.000</cell><cell>0.000</cell><cell>0.000</cell><cell>15</cell></row><row><cell>no_val</cell><cell>0.977</cell><cell>0.691</cell><cell>0.809</cell><cell>854</cell></row><row><cell>social</cell><cell>0.500</cell><cell>0.400</cell><cell>0.444</cell><cell>5</cell></row><row><cell>Accuracy Macro avg Weighted avg</cell><cell>0.430 0.809</cell><cell>0.478 0.668</cell><cell>0.668 0.425 0.704</cell><cell>1203 1203 1203</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7</head><label>7</label><figDesc>Efficiency report for the best-performing prompt (procedural_full, adapted ex-post to 3-class) on the 3-class tagset</figDesc><table><row><cell>Class</cell><cell cols="4">Precision Recall F1-score Support</cell></row><row><cell>eval_generic</cell><cell>0.553</cell><cell>0.807</cell><cell>0.657</cell><cell>244</cell></row><row><cell>eval_individual</cell><cell>0.431</cell><cell>0.629</cell><cell>0.512</cell><cell>105</cell></row><row><cell>no_val</cell><cell>0.954</cell><cell>0.775</cell><cell>0.855</cell><cell>854</cell></row><row><cell>Accuracy Macro avg Weighted avg</cell><cell>0.646 0.827</cell><cell>0.737 0.769</cell><cell>0.769 0.675 0.785</cell><cell>1203 1203 1203</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 8</head><label>8</label><figDesc>Efficiency report for the best-performing prompt (complex_binary) on the binary tagset</figDesc><table><row><cell>Class</cell><cell cols="4">Precision Recall F1-score Support</cell></row><row><cell>no_val</cell><cell>0.962</cell><cell>0.808</cell><cell>0.878</cell><cell>854</cell></row><row><cell>val</cell><cell>0.663</cell><cell>0.923</cell><cell>0.771</cell><cell>349</cell></row><row><cell>Accuracy Macro avg Weighted avg</cell><cell>0.812 0.875</cell><cell>0.865 0.841</cell><cell>0.841 0.825 0.847</cell><cell>1203 1203 1203</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://spacy.io/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://www.goodreads.com/review/show/1373835925</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://www.goodreads.com/review/show/225468198</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">The reason for the presence of such cases is not hard to find, given that, as we have said, several problems with</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">https://github.com/SimoneRebora/CHR2024</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">https://huggingface.co/GVezzani/literary_evaluation_classifier</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research was developed in the context of the "Inclusive Humanities" Excellence Project at the Department of Foreign Languages and Literatures of the University of Verona.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Power to the Reader&apos; or &apos;Degradation of Literary Taste&apos;? Professional Critics and Amazon Customers as Reviewers of the Inheritance of Loss</title>
		<author>
			<persName><forename type="first">D</forename><surname>Allington</surname></persName>
		</author>
		<idno type="DOI">10.1177/0963947016652789</idno>
	</analytic>
	<monogr>
		<title level="j">Language and Literature: International Journal of Stylistics</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="254" to="278" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Ashcroft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Grifïths</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tifïn</surname></persName>
		</author>
		<title level="m">The Empire Writes Back: Theory and Practice in Post-colonial Literatures. New accents</title>
				<meeting><address><addrLine>London; New York</addrLine></address></meeting>
		<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="1989">1989</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Les Règles de L&apos;Art: Genèse et Structure du Champ Littéraire</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bourdieu</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>du Seuil</publisher>
			<pubPlace>Paris</pubPlace>
		</imprint>
	</monogr>
	<note>Points Essais</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Unleashing the Potential of Prompt Engineering in Large Language Models: A Comprehensive Review</title>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Langrené</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="DOI">10.48550/arxiv.2310.14735</idno>
		<ptr target="https://arxiv.org/abs/2310.14735" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">18 Die Kritik und ihre Päpste: Rückblick auf ein Genre</title>
		<author>
			<persName><forename type="first">T</forename><surname>Chervel</surname></persName>
		</author>
		<idno type="DOI">10.14361/9783839454435-019</idno>
		<ptr target="https://www.transcript-open.de/doi/10.14361/9783839454435-019" />
	</analytic>
	<monogr>
		<title level="m">Digital Humanities</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Graf</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Knackstedt</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Petzold</surname></persName>
		</editor>
		<meeting><address><addrLine>Bielefeld, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>transcript Verlag</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="297" to="302" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">The Riddle of Literary Quality: A Computational Approach</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">V</forename><surname>Dalen-Oskam</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
			<publisher>Amsterdam University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Aspect-Based Sentiment Analysis for German: Analyzing &quot;Talk of Literature&quot; Surrounding Literary Prizes on Social Media</title>
		<author>
			<persName><forename type="first">L</forename><surname>De Greve</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Van Hee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Lefever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Martens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics in the Netherlands Journal</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="85" to="104" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Kinds of Literature: An Introduction to the Theory of Genres and Modes</title>
		<author>
			<persName><forename type="first">A</forename><surname>Fowler</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1985">1985</date>
			<publisher>Clarendon Press</publisher>
			<pubPlace>Oxford</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Everyone&apos;s a Critic: Rezensieren in Zeiten des Ästhetischen Plebiszit</title>
		<author>
			<persName><forename type="first">J</forename><surname>Franzen</surname></persName>
		</author>
		<idno type="DOI">10.37189/duepublico/74186</idno>
		<ptr target="https://duepublico2.uni-due.de/receive/duepublico%5C%5Fmods%5C%5F00074186" />
	</analytic>
	<monogr>
		<title level="m">Unterstellte Leseschaften: Tagung</title>
				<meeting><address><addrLine>Essen</addrLine></address></meeting>
		<imprint>
			<publisher>DuEPublico</publisher>
			<date type="published" when="2020-09-30">29. bis 30. September 2020. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The Hermeneutic Profit of Annotation: On Preventing and Fostering Disagreement in Literary Analysis</title>
		<author>
			<persName><forename type="first">E</forename><surname>Gius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jacke</surname></persName>
		</author>
		<idno type="DOI">10.3366/ijhac.2017.0194</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Humanities and Arts Computing</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="233" to="254" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Cultural Capital: The Problem of Literary Canon Formation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Guillory</surname></persName>
		</author>
		<idno type="DOI">10.7208/chicago/9780226830605.001.0001</idno>
	</analytic>
	<monogr>
		<title level="m">First edition, enlarged</title>
				<meeting><address><addrLine>Chicago</addrLine></address></meeting>
		<imprint>
			<publisher>The University of Chicago Press</publisher>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The Qualities of Literatures: A Concept of Literary Evaluation in Pluralistic Societies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Heydebrand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Winko</surname></persName>
		</author>
		<ptr target="https://www.jbe-platform.com/content/books/9789027291516-lal.4.16hey" />
	</analytic>
	<monogr>
		<title level="m">The Quality of Literature</title>
				<imprint>
			<publisher>John Benjamins</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="223" to="239" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">V</forename><surname>Heydebrand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Winko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Einführung in die Wertung von Literatur: Systematik -Geschichte -Legitimation</title>
		<title level="s">UTB für Wissenschaft</title>
		<meeting><address><addrLine>Schöningh</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
		<respStmt>
			<orgName>Uni-Taschenbücher ; Paderborn München</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A Comprehensive Review on Transformers Models For Text Classification</title>
		<author>
			<persName><forename type="first">R</forename><surname>Kora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohammed</surname></persName>
		</author>
		<idno type="DOI">10.1109/miucc58832.2023.10278387</idno>
		<ptr target="https://ieeexplore.ieee.org/document/10278387/" />
	</analytic>
	<monogr>
		<title level="m">2023 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC)</title>
				<meeting><address><addrLine>Cairo, Egypt</addrLine></address></meeting>
		<imprint>
			<publisher>Ieee</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Absorption in Online Reviews of Books: Presenting the English-Language AbsORB Metadata Corpus and Annotation Guidelines</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kuijpers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lendvai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lusetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rebora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ruh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tadres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ternes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vogelsanger</surname></persName>
		</author>
		<idno type="DOI">10.5334/johd.116</idno>
		<ptr target="http://openhumanitiesdata.metajnl.com/articles/10.5334/johd.116/" />
	</analytic>
	<monogr>
		<title level="j">Journal of Open Humanities Data</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page">13</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Detection of Reading Absorption in User-Generated Book Reviews: Resources Creation and Evaluation</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lendvai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Darányi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Geng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kuijpers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Lopez De Lacalle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-C</forename><surname>Mensonides</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rebora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Reichel</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.lrec-1.595" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twelfth Language Resources and Evaluation Conference</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Mazo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Moreno</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Twelfth Language Resources and Evaluation Conference<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<publisher>European Language Resources Association</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="4835" to="4841" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Birds of a Feather Sing Together</title>
		<author>
			<persName><forename type="first">N</forename><surname>Mark</surname></persName>
		</author>
		<idno type="DOI">10.2307/3005535</idno>
	</analytic>
	<monogr>
		<title level="j">Social Forces</title>
		<imprint>
			<biblScope unit="volume">77</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">453</biblScope>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Interrater Reliability: The Kappa Statistic</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Mchugh</surname></persName>
		</author>
		<idno type="DOI">10.11613/bm.2012.031</idno>
	</analytic>
	<monogr>
		<title level="j">Biochemia Medica</title>
		<imprint>
			<biblScope unit="page" from="276" to="282" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Pianzola</surname></persName>
		</author>
		<title level="m">Digital Social Reading: Sharing Fiction in the Twenty-First Century</title>
				<meeting><address><addrLine>Cambridge, Massachusetts</addrLine></address></meeting>
		<imprint>
			<publisher>The MIT Press</publisher>
			<date type="published" when="2025">2025</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Towards Hermeneutic Markup: An architectural outline</title>
		<author>
			<persName><forename type="first">W</forename><surname>Piez</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:15829935" />
	</analytic>
	<monogr>
		<title level="m">Digital Humanities Conference</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">GPT Is an Effective Tool for Multilingual Psychological Text Analysis</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rathje</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D.-M</forename><surname>Mirea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sucholutsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Marjieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J V</forename><surname>Bavel</surname></persName>
		</author>
		<idno type="DOI">10.31234/osf.io/sekf5</idno>
		<ptr target="https://osf.io/sekf5" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Digital Humanities and Digital Social Reading</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rebora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Boot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pianzola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Gasser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Herrmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kraxenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Kuijpers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lendvai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">C</forename><surname>Messerli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sorrentino</surname></persName>
		</author>
		<idno type="DOI">10.1093/llc/fqab020</idno>
	</analytic>
	<monogr>
		<title level="j">Digital Scholarship in the Humanities</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">Supplement_2</biblScope>
			<biblScope unit="page" from="230" to="250" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Mining Goodreads. A Digital Humanities Project for the Study of Reading Absorption</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rebora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kuijpers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lendvai</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.3897251</idno>
		<ptr target="https://zenodo.org/record/3897251" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Comparing ChatGPT to Human Raters and Sentiment Analysis Tools for German Children&apos;s Literature</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rebora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Lehmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Heumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lauer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Humanities Research Conference</title>
				<imprint>
			<date type="published" when="2023">CHR2023. 2023</date>
			<biblScope unit="page" from="333" to="343" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Orientalism</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">W</forename><surname>Said</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1978">1978</date>
			<publisher>Pantheon Books</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
	<note>1st ed</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Literary Value in the Era of Big Data. Operationalizing Critical Distance in Professional and Non-professional Reviews</title>
		<author>
			<persName><forename type="first">M</forename><surname>Salgaro</surname></persName>
		</author>
		<idno type="DOI">10.22148/001c.36446</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Cultural Analytics</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">How to Measure the Social Prestige of a Nobel Prize in Literature? Development of a Scale Assessing the Literary Value of a Text</title>
		<author>
			<persName><forename type="first">M</forename><surname>Salgaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sorrentino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gerhard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jacobs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Txt</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="138" to="148" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">How a Literacy Work Becomes a Masterpiece: On the Threefold Selection Practised by Literary Criticism</title>
		<author>
			<persName><forename type="first">C</forename><surname>Van Rees</surname></persName>
		</author>
		<idno type="DOI">10.1016/0304-422x(83)90015-3</idno>
	</analytic>
	<monogr>
		<title level="j">Poetics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">4-5</biblScope>
			<biblScope unit="page" from="397" to="417" />
			<date type="published" when="1983">1983</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">The Legitimacy of Book Critics in the Age of the Internet and Omnivorousness: Expert Critics, Internet Critics and Peer Critics in Flanders and the Netherlands</title>
		<author>
			<persName><forename type="first">M</forename><surname>Verboord</surname></persName>
		</author>
		<idno type="DOI">10.1093/esr/jcp039</idno>
	</analytic>
	<monogr>
		<title level="j">European Sociological Review</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="623" to="637" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Exploring Goodreads Reviews for Book Impact Assessment</title>
		<author>
			<persName><forename type="first">K</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Han</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.joi.2019.07.003</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Informetrics</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="874" to="886" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">A</forename></persName>
		</author>
		<title level="m">The Tag Set Here&apos;s a list of the labels in our tag set</title>
				<imprint/>
	</monogr>
	<note>each followed by a brief description</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
