=Paper= {{Paper |id=Vol-3834/paper124 |storemode=property |title=Once More, With Feeling: Measuring Emotion of Acting Performances in Contemporary American Film |pdfUrl=https://ceur-ws.org/Vol-3834/paper124.pdf |volume=Vol-3834 |authors=Naitian Zhou,David Bamman |dblpUrl=https://dblp.org/rec/conf/chr/ZhouB24 }} ==Once More, With Feeling: Measuring Emotion of Acting Performances in Contemporary American Film== https://ceur-ws.org/Vol-3834/paper124.pdf
                                Once More, With Feeling: Measuring Emotion of
                                Acting Performances in Contemporary American
                                Film
                                Naitian Zhou∗ , David Bamman
                                School of Information, UC Berkeley, USA


                                           Abstract
                                           Narrative film is a composition of writing, cinematography, editing, and performance. While much
                                           computational work has focused on the writing or visual style in film, we conduct in this paper a com-
                                           putational exploration of acting performance. Applying speech emotion recognition models and a vari-
                                           ationist sociolinguistic analytical framework to a corpus of popular, contemporary American film, we
                                           find narrative structure, diachronic shifts, and genre- and dialogue-based constraints located in spoken
                                           performances.

                                           Keywords
                                           film, performance, computational film analysis, speech emotion recognition




                                1. Introduction
                                Film is rich in its supply of semiotic resources, communicating meaning from the interaction
                                of language (encoded in a script), visuals (choices of composition, blocking, cinematography),
                                sound and more. Much computational work has arisen to examine slices of this semiotic field,
                                including measuring how gender stereotypes or plot arcs are reflected in dialogue [34, 29, 14] or
                                how visual features like color variance and shot length constitute genre [25, 11]. One critical
                                area, however, that has been neglected in this study is the role of performance in creating
                                meaning.
                                   As Naremore [18] notes, film is a medium in which meaning is acted out; an acting per-
                                formance provides a semiotic frame through which we can understand the events that unfold.
                                Given the fixed text of a script, the rendering of the final performance is an interpretive process
                                in which the actor, director and editor jointly imbue the words with additional meaning. In
                                this view, the same line of dialogue exhibits variation in meaning when performed in distinct
                                diegetic contexts. As one example, consider the following line in Knives Out (2019):

                                                                                             “I’m warning you.”

                                  Much of the film revolves around these three words, overheard in a conversation between
                                the wealthy Harlan Thrombey and his grandson, Ransom. The line is uttered by multiple char-
                                CHR 2024: Computational Humanities Research Conference, December 4–6, 2024, Aarhus, Denmark
                                ∗
                                 Corresponding author.
                                £ naitian@berkeley.edu (N. Zhou); dbamman@berkeley.edu (D. Bamman)
                                ȉ 0009-0005-1991-2258 (N. Zhou)
                                         © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).




                                                                                                          186
CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
acters as the film unfolds: angrily shouted by Ransom, somberly recalled by the eavesdropper,
and gleefully recounted by inspector Benoit Blanc upon solving the crime. Even a single line,
located within a single diegetic event, has great capacity for meaning-making in performance.
   When viewed in this light, we can apply the analytical framework from variationist soci-
olinguistics to better understand this space of performance. Given a fixed line of dialogue
(equivalent to a linguistic variable), a performance entails a choice — a selection from the set
of possible variants. It is this choice, and the meaning contained within, which we study.
   In this work, we design computational models to explore this form of variation by consider-
ing the emotional range of performances in contemporary American film, exploring in particu-
lar the tension between what characters say and how they say it. As distinct from prior work in
the computational humanities that has measured emotion from text alone [14, 15], we measure
acted emotion from speech, allowing us to disentangle the emotion present in the script from
the choices made in creating the performance.
   Using a speech emotion recognition model, we construct a parallel dataset of spoken per-
formances (utterances) aligned with the text of the words being spoken (dialogue phrases).
This dataset allows us to isolate and examine how performances vary in meaning from their
paralinguistic features in addition to the textual meaning of the screenplay. We use this dataset
to carry out several case studies exploring variation in performance in American film:
   1. First, we carry out a structural analysis of emotion as performed over narrative time. Do-
      ing so allows us to characterize film as performance text, relating emotional performance
      to larger narrative structure.
   2. Second, we study diachronic variation by comparing emotionality of films across release
      years, testing the degree to which performances have intensified over time (following
      Bordwell’s theories of visual style [4]).
   3. Finally, we examine the capacity for performance by constructing a novel measure of
      emotional range for an utterance—the space of possible emotions that can be performed.
      In doing so, we demonstrate how both contextual (genre) and textual (dialogue) aspects
      of film can carry constraints and affordances on acting performance.
  In this work, we use computational methods to survey how both textual and contextual
variables inform and reflect the performances rendered on screen.


2. Methods
In order to perform our analysis, we need to construct an aligned dataset of actor performances
(utterances), the text of the words they speak (phrases), and the emotions in each utterance.
We create a pipeline that takes as input a set of full-length movies, and outputs time-aligned
transcriptions for utterances, their emotion labels, and groups of semantically similar phrases.

2.1. Preprocessing pipeline
We first construct a data pipeline to segment and transcribe utterances from movie dialogues.
The pipeline takes as input a set of MP4 files, where each file is one digitized film. Our analysis
takes place in the speech and text modalities, so we use ffmpeg to extract the audio track.




                                               187
   We use the pyannote1 segmentation model to detect continuous, single-speaker speech seg-
ments. Then, we use faster-whisper2 to transcribe each speech segment. Because pyannote
speech segments are based on voice activity detection and silences, it can label extended, multi-
sentence speech as a single segment. For our analysis, however, we are interested in utterances
as a discursive unit. If a character makes an assessment, then poses a question, we would like
to split these into two distinct utterances. As a middle ground between raw voice activity detec-
tion and segmenting discursive units, which requires complex conversational understanding,
we perform a post-processing step where we further split speech segments by sentence bound-
aries derived from the transcriptions. Because whisper is an end-to-end model that does not
produce fine-grained time alignments, we then use a speech-to-text fine-tuned wav2vec23 to
perform word-level time alignment between the transcription and the audio, then split the au-
dio based on sentence boundaries generated by syntok,4 a fast, rule-based sentence segmenter.
   To prevent the end credit sequences from interfering with the results, we detect when the
end credits begin by performing optical character recognition (OCR) on the shots in a movie
and identifying long continuous sequences of shots that contain large amounts of text. We
trim the movie to the beginning of the end credits.

2.2. Speech emotion recognition
To perform speech emotion recognition, we use a wav2vec2 large model without any task-
specific fine-tuning to extract audio features. Then, we train a classification head to perform
seven-way emotion classification, based on the Ekman model [9] of six basic emotions (anger,
disgust, happiness, sadness, fear, and surprise) and a neutral label. To train these models, we
use the MELD dataset, which contains 1,000 sampled dialogues from the TV series Friends [24].
   We experiment with two classification settings: an utterance-level model which makes pre-
dictions based on only the speech features of the input utterance and a conversation-level model
which includes the speech features of both the input utterance and its surrounding utterances.
   In both cases, we use a pretrained wav2vec2 model as a backbone model for generating vector
representations of each utterance. Because wav2vec2 creates an embedding for each audio
frame (roughly 20ms of speech), we follow prior work in computing utterance embeddings by
averaging across all timestamps within an utterance [21]. We compute embeddings from the
attention activations at each layer of the wav2vec2 model instead of just taking the last-layer
activations; prior work has shown that, for paralinguistic tasks such as emotion recognition,
early- and intermediate-layer activations are more useful than later layers [21, 30]. At the end
of the embedding step, each utterance is represented by a set of 25 768-dimensional vectors.

2.2.1. Utterance-level emotion recognition
We implement the utterance-level model from Pepino, Riera, and Ferrer [21] and match the re-
ported performance. We first take a weighted average of layer activations for a given utterance;

1
  https://huggingface.co/pyannote/speaker-segmentation
2
  https://github.com/SYSTRAN/faster-whisper
3
  https://huggingface.co/facebook/wav2vec2-base-960h
4
  https://github.com/fnl/syntok




                                                    188
these weights are learned during training. Then, we apply a fully-connected classification head
to produce a probability distribution over the seven emotion labels. Unlike the original paper,
we do not use features from the initial convolutional layer of the pre-trained model; we use
only the attention head activations.

2.2.2. Contextual emotion recognition
We also train an contextual model which uses a bidirectional LSTM to predict the emotion of
utterances within the context of a conversation. To do so, we define conversations as groups
of utterances where each occurs within 3 seconds of the next. In the MELD dataset, there are
1,478 conversations in the training split according to this criterion.
   For each conversation, we predict the emotion labels of all utterances in the conversation by
first passing weighted activations through the biLSTM before applying the classification head
to each hidden state. As before, the weights of activations are learned during training.

2.2.3. Evaluation
We expect the movie data to be similar in nature to the MELD dataset, since both consist of
professionally produced and acted clips. However, to ensure that our models do not experience
domain shift despite the greater range in release year and setting of the film corpus, we evaluate
these models on the test split of the MELD dataset as well as a manually collected dataset
consisting of 333 clips from a subset of 35 contemporary American films. Each clip was a
conversation with at least 2 utterances, where conversations were identified with the same
heuristic used to construct training data for the contextual model. This resulted in a final
evaluation dataset of 2,157 utterances with emotion labels. The clips were labeled by two
annotators: 51 clips were labeled by both annotators and 250 clips were labeled by a single
annotator. The Krippendorff’s 𝛼 between the two annotators was 0.334, and the Fleiss’ 𝜅 was
0.333, which matches the agreement of the MELD dataset.
   Table 1 shows the evaluation results on the MELD and Movies datasets. The models perform
comparably to each other, and comparably across evaluation datasets. This performance also
approaches the state of the art on MELD for audio-only models. Because the performance of
the contextual model is slightly higher, we use its inference outputs for our analysis.

Table 1
Model comparison for MELD test dataset and our movies dataset, along with 95% bootstrap confidence
interval bounds.
                                MELD                                          Movies
                    Accuracy             Weighted F1              Accuracy             Weighted F1
 Utterance     0.476 [0.451, 0.501]   0.455 [0.432, 0.478]   0.449 [0.429, 0.469]   0.434 [0.413, 0.456]
 Contextual    0.494 [0.472, 0.516]   0.456 [0.441, 0.472]   0.488 [0.468, 0.508]   0.450 [0.428, 0.472]




                                                  189
2.3. Identifying dialogue phrase groups
One powerful aspect of this dataset is that we align actor performances to the words that they
speak. To account for variation in how highly semantically similar phrases can be realized, we
cluster together phrases with high semantic similarity. We use the sentence-transformers
library to compute sentence embeddings of utterances and cluster them with the Leiden com-
munity detection algorithm [31]. Table 2 shows some examples of phrases that are grouped
together. We expect the phrases in each group to have similar prior distributions of emotion.

Table 2
Examples of utterances which are clustered into dialogue phrase groups.

         Phrase Groups
         “Let’s go, let’s go, let’s go!”, “Let’s go, let’s go!”, “Let’s go right now go go”, “Go,
         let’s go, let’s go.”, “Okay guys, let’s go.”
         “Oh, pleasure to meet you.”, “It’s so nice to finally meet you.”, “It is a pleasure to
         finally meet you.”, “Oh, it’s nice to meet you.”, “It’s so nice to meet you!”




3. Analysis
The above pipeline measures the emotions performed in an utterance and ties each utterance to
the text being spoken. We apply this methodology to a large corpus of contemporary, popular
American films [1] in order to study the variation of emotion within and between them.
   Our corpus consists of the top-50 live-action, narrative films by U.S. box ofÏce from 1980-
2022. We supplement these with films nominated for “Best Picture”-equivalent awards by one
of six organizations in those years: Academy Awards, Golden Globes, British Academy of Film
and Television Arts, Los Angeles Film Critics Association, National Board of Review, and Na-
tional Society of Film Critics. We only include English-language films in this analysis, resulting
in a total of 2,283 films.

3.1. Film as performance text
Plantinga [22] describes how emotionality can reflect narrative structure — emotionally-
charged events can serve as catalyst to disrupt the expository “stable state” and set the narrative
in motion. Much attention has been paid to characterizing narratives in literature and film in
terms of emotionality using trajectories of sentiment [13] and emotion [26, 14, 12, 32]; these
have focused on the emotion encoded in text. Audiences of movies, however, are not directly
exposed to that text; their experience is mediated by the performance. In order to study this
more directly, we turn our attention to characterizing narratives with emotion as performed.
   We study the distribution of emotions in utterances over the course of a movie. How do
the prevalence of emotions shift over narrative time? Similar to previous work on dialogue
in screenplays, we ask if there are emotional regularities across films [12]. We examine first
the emotionality of utterances — the average probability that an utterance is not neutral —




                                                   190
                0.50

                                                                           0.48
                0.48
     Emotionality



                                                                           0.46
                0.46




                                                                    Joy
                                                                           0.44
                0.44

                                                                           0.42
                0.42

                       10% 20% 30% 40% 50% 60% 70% 80% 90% 100%                   10% 20% 30% 40% 50% 60% 70% 80% 90% 100%
                                    Narrative Time                                             Narrative Time

                               (a) Emotionality                                                 (b) Joy

                                                                       0.155
             0.155
             0.150                                                     0.150

             0.145                                                     0.145
             0.140
    Sadness




                                                                       0.140
                                                                   Anger

             0.135
             0.130                                                     0.135

             0.125                                                     0.130
             0.120
                                                                       0.125
                       10% 20% 30% 40% 50% 60% 70% 80% 90% 100%                   10% 20% 30% 40% 50% 60% 70% 80% 90% 100%
                                    Narrative Time                                             Narrative Time

                                  (c) Sadness                                                 (d) Anger
Figure 1: Emotionality increases, but specific emotions show non-linear trajectories over narrative time
(95% bootstrap confidence interval).


before looking more closely at how specific emotions are distributed temporally. We plot the
average probability of an emotion label for an utterance in intervals of 5 percent, expressed as
a percentage of the full run-time of the film. Specific emotions are measured as proportions of
the emotional labels, excluding the neutral label.
   We find that the emotional trajectories of performances are, in fact, structured over narrative
time. Figure 1a shows that emotionality increases over narrative time. We examine also the
trajectory of specific emotions across films (figs. 1b,1c,1d). We find that joyful performances
follow a U-shaped curve, with a steep increase towards the end, as movies resolve. Like Hipson
and Mohammad [12], we find that negative-valence emotions like sadness and anger decrease
at the end. Further, anger peaks 85% into the film, reminiscent of a climax-resolution structure.

3.2. Evolving emotionality
Subscribing to a particular categorization of emotions can be restrictive; in the remainder of
the paper, we explore emotional performance, but depart from analyzing specific emotional
labels. First, we study performance at a less granular level, focusing on the concept of emotion-




                                                                  191
                                       0.50

                                       0.48
                    Average Emotionality

                                       0.46

                                       0.44

                                       0.42


                                              1980   1990     2000         2010   2020
                                                            Release Year
Figure 2: Emotionality is higher in older films (95% bootstrap confidence interval).


ality as the proportion of utterances with any emotion label.5 We measure how emotionality
has changed historically over the decades spanned by our corpus. Emotional shifts have been
identified in English fiction books: Morin and Acerbi [17] find that the content of those stories
have experienced a decline in emotional expression. Within cinema, David Bordwell has writ-
ten about how shorter shot lengths and tighter framing serve to intensify the visual style in
more recent films compared to earlier ones [4]. We ask whether there is a similar shift in per-
formance: is there an intensification of emotion that matches the visual intensification of film,
or perhaps an emotional cooling in performance that matches the findings in English fiction?
   When we split the data by release year, we find a mild effect that earlier films have a higher
proportion of emotional utterances compared to later ones, with emotionality hitting a mini-
mum around 2010 (see Fig. 2). However, the question remains whether the emotional content
is changing (as Morin and Acerbi find in literature) or if the style with which words are being
uttered is changing.
   To disentangle the effects of shifting content and shifting style, we consider the change in
emotionality over the years within the semantically equivalent phrase groups. If it is indeed
the writing, and not the performance, that drives this shift in emotionality, we should see little
change within a phrase group. However, when we look at the 511 phrases that are used in all
43 years of the dataset, we find that a fixed-effects regression shows a slightly negative, sta-
tistically significant correlation between the year and emotionality even within phrase groups
(𝑅2 = 0.048, 𝐹 (1, 21461), 𝑝 < 0.001).
   Though this result is seemingly at odds with Bordwell’s finding that visual style intensifies, it
is also possible that they are harmonious. In Hollywood film, the close-up shot has always been

5
    The model achieves an F1 score of 0.69 on the neutral label.




                                                               192
associated with emotional expression [20]. Panovsky [19] writes that close-ups provide a rich
“field of action” that affords nuanced acting performances. These visual performances, which
are almost imperceptible if viewed from a natural distance, provide an alternative to the spoken
word as a channel of expression. Comparing to the stage, Panovsky writes the spoken word
makes a stronger impression “if we are not permitted to count the hairs in Romeo’s mustache.”
As cinema further grows into its medium, Bordwell finds that close-up shots have indeed grown
tighter on the subject. With an increase in the capacity for more nuanced performance in the
visual channel, the emotionality of the spoken word need not bear so strong a burden.

3.3. Measuring emotional range
Range is often said to be the mark of a great actor. Naremore writes about the importance of an
actor splitting their character “visibly into different aspects”, showing off emotional range [18].
Kuleshov similarly stressed the actors must be able to create a full range of gestures to create
complex meaning [16]. Wilson [33] takes this a step further and argues that the hallmark of
great acting is projecting a character into complex situations. In this section, we explore the
limits of range through the constraints that genre and script impose on emotional performance.
   For this analysis, we construct a general measure of emotional range across a set of utter-
ances 𝑢1…𝑛 . We characterize each utterance 𝑢𝑖 with a performance vector 𝑣⃗𝑖 , which is a distri-
bution over emotions, given by the predicted probability distribution from the speech emotion
recognition model. This allows us to take a more nuanced view of performances as a mixture of
emotions. We model the distribution from which the vectors 𝑣⃗1…𝑛 are drawn as a Dirichlet, and
find the parameters which maximize the likelihood of the observed vectors. We define emo-
tional range as the entropy of this distribution: a higher entropy means there is greater variance
in the distribution of performances, and a lower entropy signals lower emotional range.
   One criticism of the Ekman emotional model lies in its construct validity: seven discrete
emotion labels may be insufÏcient to characterize the space of emotions. Ideally, we would
model a continuous space of “performance”. In our previous analyses, we use these emotion
labels as an intermediate between that ideal on one end, and sentiment analysis on the other.
Here, our measure of emotional range is agnostic to the meaning of specific emotion labels, and
serves to demonstrate how emotion classification can be a useful proxy task through which we
can analyze performance in a more continuous space.

Thrillers have the least range; family-friendly films have the most. Wilson [33] specu-
lates that some genres, like some types of comedy, have less capacity for emotional range than
others. Previous work has shown that emotional arcs are correlated with genre [28]. We ask
whether different genres are associated with different capacities for emotional performance.
   We calculate the emotional range for each movie, and find the average score for each genre.
Genre information comes from IMDB, and we exclude genres with fewer than 30 films in our
dataset.6 Figure 3 shows the average entropy across genres. Thrillers, biographies, and mys-
teries have the least emotional range; fantasy, musicals and family films rank highest. While


6
    Three genres were excluded: Western (19 films), Documentary (3), and Animation (2)




                                                       193
                      Thriller
                   Biography
                     Mystery
                       Crime
                        Sci-Fi
                       Action
                    Romance
                      History
                      Drama
               Genre

                        Music
                     Comedy
                        Sport
                         War
                   Adventure
                       Horror
                     Fantasy
                      Musical
                       Family
                                 11.4 11.2 11.0 10.8 10.6 10.4 10.2 10.0         9.8
                                                 Emotional Range

Figure 3: Relative emotional range for different film genres (95% bootstrap confidence intervals).


it is difÏcult to attribute these results to a particular property of specific genres, these findings
show that some genres have more constrained or consistent emotional registers than others.

Functional phrases have less capacity for emotional range. Naremore [18] references
Goffman when theorizing about performance: actors draw on and play against the interactional
norms with which we as audience are already familiar. We ask if this bears out in our data. Does
the emotional range of dialogue phrases reflect their discursive properties?
   To study this, we measure the emotional range in dialogue. Because we tie specific perfor-
mances to the words that are spoken, we can identify instances across the corpus when a given
phrase was uttered. We isolate the 2,656 phrase groups that are uttered at least 50 times across
our dataset. For each phrase group, we calculate the emotional range of its utterances.
   Table 3 shows phrases with the highest and lowest entropies. By inspecting the phrases at
either end of the spectrum, we find qualitative differences in the kinds of phrases that have
higher and lower emotional range: the capacity for emotional variance reflects the discursive
flexibility of the words being spoken. Phrases with low range are functional and generally
part of highly directed interactions: most phrases are either yes-or-no questions or answers to
them. Phrases with high emotional range, on the other hand, mostly have more open-ended,
evaluative discursive functions. In these cases, the prosody or intonation of speech can easily
lend color to the statement being made. “You’re alive” can be said with joy or relief to a loved
one, as Marty McFly to his mentor Doc in Back to the Future (1985), or with anger at the sight
of an enemy, as Lord Norinaga greets Walker in Teenage Mutant Ninja Turtles III (1993).




                                                 194
Table 3
Dialogue phrase groups with the highest and lowest emotional range scores. The table shows a repre-
sentative phrase from each group.

             Low Emotional Range                             High Emotional Range
  Phrase                               Entropy     Phrase                             Entropy
  “Could I ask you something?”          -17.02     “All rise.”                          -7.85
  “This is your captain speaking.”      -16.56     “Are you out of your mind?”          -7.88
  “Is that okay?”                       -16.53     “What the fuck wrong with you?”      -7.99
  “Can I get something for you?”        -16.17     “You’re alive.”                      -8.32
  “Can I get something to drink?”       -16.16     “You saved my life.”                 -8.34
  “Hey, what can I get you?”            -15.91     “Don’t you understand?”              -8.34
  “You wanna come?”                     -15.65     “Don’t be so afraid.”                -8.39
  “Yeah, that’s good.”                  -15.36     “You son of a bitch.”                -8.40
  “Any questions?”                      -15.26     “You scared the shit out of me!”     -8.44
  “That’s correct.”                     -15.25     “Ow.”                                -8.44


4. Discussion and limitations
With this work, we demonstrate that films can, and should, be studied as performance texts.
We tie our findings to both film theory and other computational work on narratives. Here, we
discuss some limitations of the current study.

Measuring emotions. We follow a vast body of previous work within natural language
processing [36, 35], affective computing [5, 6] and computational literary studies [28, 3] in
using Ekman’s basic emotions. However, the validity of this model has been questioned [23].
   First, there are doubts about the ecological validity of emotion recognition, especially as most
speech emotion recognition datasets contain acted emotion as opposed to natural emotion. We
note that, unlike much affective computing work, we use emotion recognition models trained
on acted speech to make inference on acted speech. The professionally-produced, acted speech
in the MELD dataset is well-suited to our data, which is also professionally-produced and acted.
Indeed, we find that performance is similar between MELD and our in-domain evaluation data.
   Another criticism lies in the cultural relativity of emotion. Though Ekman argues that the
basic emotions are universal, he acknowledges there may be cultural differences in the emo-
tions elicited in a given context. It is reasonable to suppose that viewers’ normative knowledge
also influences the interpretation of these emotions. We focus on contemporary American film
in both our analysis and training data, holding at least the intended cultural audience constant.
Cultural variation in performance is a ripe area for future work, as cultural differences exist in
not only the production and interpretation of emotion, but also in theories of acting.
   Aside from these specific criticisms of the Ekman model, the low interannotator agreement
in both our evaluation set as well as other datasets, including MELD, suggest that this model
for emotion may remain too coarse to precisely describe the data. Work in both affective psy-




                                                 195
chology and NLP have attempted to address this by using more fine-grained classes [8, 7] or a
continuous spaces of emotion [27, 7]. While we used the Ekman model due to the availability
of training data as well as to provide comparison with previous studies of emotion narratives,
alternative emotion models may prove useful in future work.

A question of authorship. In cinema, the performance that audiences see on screen is co-
created by the actor, the director, and the editor. Baron and Carnicke [2] describe the conven-
tional wisdom within film analysis to be that cinematic performances are made in the cutting
room. “True” acting happens on the stage. Though our work studies film as performance text,
it does not disentangle the processes through which the performance is constructed. It is about
the performance as viewed, but not about the choices made by actors as separate from the di-
rector or editor. Our work makes the point that performance carries meaning worth studying,
and opens the door for future computational work that explores its authorial roots.

Embodied erformance. Finally, we examine only performance as enacted through speech.
This is perhaps the modality that lies closest to the script, and allows us to apply a variationist
approach to studying the relationship between performance and text, but of course perfor-
mance includes not just speech but also gesture, posture, facial expression, and more. Visual
description has been found to be more useful for aligning narrative events than dialogue [37],
and quantitative analysis of theater performance has found narratively meaningful patterns in
movement [10]. Film is a multimodal medium that deserves analysis in all its modalities. We
hope that our work examining film across the speech and text can serve as a basis for more
work that examines performance as embodied visually.


5. Conclusion
In this paper, we explore the relation between film as narrative text and as performance text.
Using a novel parallel dataset of speech and text from popular contemporary American film,
we develop computational methods to measure how emotional prevalence and emotional range
vary by both textual factors of narrative time and dialogue, as well as contextual factors of
release year and genre. We hope this work inspires further multimodal studies of performance
in computational film analysis.


Acknowledgments
The research reported in this article was supported by funding from Mellon Foundation and
the National Science Foundation (IIS-1942591 and DGE-2146752). We thank Jacob Lusk and
Lucy Li for insightful discussion and feedback.




                                               196
References
 [1] D. Bamman, R. Samberg, R. J. So, and N. Zhou. “Measuring Diversity in Hollywood
     through the Large-Scale Computational Analysis of Film”. In: Proceedings of the National
     Academy of Sciences (2024).
 [2] C. Baron and S. Carnicke. Reframing Screen Performance. Ann Arbor, MI: University of
     Michigan Press, 2008. doi: 10.3998/mpub.104480.
 [3] L. Barros, P. Rodriguez, and A. Ortigosa. “Automatic Classification of Literature Pieces
     by Emotion Detection: A Study on Quevedo’s Poetry”. In: 2013 Humaine Association Con-
     ference on Affective Computing and Intelligent Interaction. 2013, pp. 141–146. doi: 10.110
     9/acii.2013.30.
 [4] D. Bordwell. “Intensified Continuity Visual Style in Contemporary American Film”. In:
     Film Quarterly 55.3 (2002), pp. 16–28. doi: 10.1525/fq.2002.55.3.16.
 [5] C. Busso, M. Bulut, C.-C. Lee, A. Kazemzadeh, E. Mower, S. Kim, J. N. Chang, S. Lee, and
     S. S. Narayanan. “IEMOCAP: Interactive Emotional Dyadic Motion Capture Database”.
     In: Language Resources and Evaluation 42.4 (2008), pp. 335–359. doi: 10.1007/s10579-008
     -9076-6.
 [6] H. Cao, D. G. Cooper, M. K. Keutmann, R. C. Gur, A. Nenkova, and R. Verma. “CREMA-D:
     Crowd-sourced Emotional Multimodal Actors Dataset”. In: IEEE transactions on affective
     computing 5.4 (2014), pp. 377–390. doi: 10.1109/taffc.2014.2336244.
 [7] A. Cowen, D. Sauter, J. L. Tracy, and D. Keltner. “Mapping the Passions: Toward a High-
     Dimensional Taxonomy of Emotional Experience and Expression”. In: Psychological Sci-
     ence in the Public Interest 20.1 (2019), pp. 69–90. doi: 10.1177/1529100619850176.
 [8] D. Demszky, D. Movshovitz-Attias, J. Ko, A. Cowen, G. Nemade, and S. Ravi. GoEmotions:
     A Dataset of Fine-Grained Emotions. 2020. arXiv: 2005.00547 [cs].
 [9] P. Ekman. “An Argument for Basic Emotions”. In: Cognition and Emotion 6.3-4 (1992),
     pp. 169–200. doi: 10.1080/02699939208411068.
[10]   M. Escobar Varela and G. O. F. Parikesit. “A Quantitative Close Analysis of a Theatre
       Video Recording”. In: Digital Scholarship in the Humanities 32.2 (2017), pp. 276–283. doi:
       10.1093/llc/fqv069.
[11]   T. Guha, N. Kumar, S. S. Narayanan, and S. L. Smith. “Computationally Deconstructing
       Movie Narratives: An Informatics Approach”. In: 2015 IEEE International Conference on
       Acoustics, Speech and Signal Processing (ICASSP). 2015, pp. 2264–2268. doi: 10.1109/icass
       p.2015.7178374.
[12]   W. E. Hipson and S. M. Mohammad. “Emotion Dynamics in Movie Dialogues”. In: Plos
       One 16.9 (2021), e0256153. doi: 10.1371/journal.pone.0256153.
[13]   M. L. Jockers. “Syuzhet: Extract Sentiment and Plot Arcs from Text”. In: (2015).




                                              197
[14]   A. K. Kayhani, F. Meziane, and R. Chiky. “Movies Emotional Analysis Using Textual
       Contents”. In: Natural Language Processing and Information Systems. Ed. by E. Métais, F.
       Meziane, H. Horacek, and P. Cimiano. Vol. 12089. Cham: Springer International Publish-
       ing, 2020, pp. 205–212. doi: 10.1007/978-3-030-51310-8\_19.
[15]   E. Kim, S. Padó, and R. Klinger. “Investigating the Relationship between Literary Gen-
       res and Emotional Plot Development”. In: Proceedings of the Joint SIGHUM Workshop
       on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Liter-
       ature. Vancouver, Canada: Association for Computational Linguistics, 2017, pp. 17–26.
       doi: 10.18653/v1/W17-2203.
[16]   L. V. Kuleshov. Kuleshov on Film: Writings. Berkeley : University of California Press, 1974.
[17]   O. Morin and A. Acerbi. “Birth of the Cool: A Two-Centuries Decline in Emotional Ex-
       pression in Anglophone Fiction”. In: Cognition & Emotion 31.8 (2017), pp. 1663–1675. doi:
       10.1080/02699931.2016.1260528.
[18]   J. Naremore. Acting in the Cinema. University of California Press, 1988.
[19]   E. Panovsky. “Style and Medium in the Moving Pictures”. In: Film, an Anthology / Com-
       piled and Edited by Daniel Talbot. New York: Simon and Schuster, 1959.
[20]   K. Pendlebury. “Cutting across the Century: An Investigation of the Close up and the
       Long-Shot in “Cine Choreography” since the Invention of the Camera”. In: The Interna-
       tional Journal of Screendance 4 (2014). doi: 10.18061/ijsd.v4i0.4527.
[21]   L. Pepino, P. Riera, and L. Ferrer. Emotion Recognition from Speech Using Wav2vec 2.0
       Embeddings. 2021. arXiv: 2104.03502 [cs, eess].
[22]   C. R. Plantinga. Moving Viewers: American Film and the Spectator’s Experience. Berkeley:
       University of California Press, 2009.
[23]   F. M. Plaza-del-Arco, A. Curry, A. C. Curry, and D. Hovy. Emotion Analysis in NLP: Trends,
       Gaps and Roadmap for Future Directions. 2024. arXiv: 2403.01222 [cs].
[24]   S. Poria, D. Hazarika, N. Majumder, G. Naik, E. Cambria, and R. Mihalcea. MELD: A
       Multimodal Multi-Party Dataset for Emotion Recognition in Conversations. 2019. arXiv:
       1810.02508 [cs].
[25]   Z. Rasheed, Y. Sheikh, and M. Shah. “On the Use of Computable Features for Film Classi-
       fication”. In: IEEE Transactions on Circuits and Systems for Video Technology 15.1 (2005),
       pp. 52–64. doi: 10.1109/tcsvt.2004.839993.
[26]   A. J. Reagan, L. Mitchell, D. Kiley, C. M. Danforth, and P. S. Dodds. “The Emotional Arcs
       of Stories Are Dominated by Six Basic Shapes”. In: EPJ Data Science 5.1 (2016), pp. 1–12.
       doi: 10.1140/epjds/s13688-016-0093-1.
[27]   J. A. Russell. “A Circumplex Model of Affect.” In: Journal of Personality and Social Psy-
       chology 39.6 (1980), pp. 1161–1178. doi: 10.1037/h0077714.
[28]   S. Samothrakis and M. Fasli. “Emotional Sentence Annotation Helps Predict Fiction
       Genre”. In: Plos One 10.11 (2015), e0141922. doi: 10.1371/journal.pone.0141922.




                                               198
[29]   B. M. Schmidt. “Plot Arceology: A Vector-Space Model of Narrative Structure”. In: 2015
       IEEE International Conference on Big Data (Big Data). 2015, pp. 1667–1672. doi: 10.1109
       /BigData.2015.7363937.
[30]   J. Shor and S. Venugopalan. “TRILLsson: Distilled Universal Paralinguistic Speech Repre-
       sentations”. In: Interspeech 2022. 2022, pp. 356–360. doi: 10.21437/Interspeech.2022-118.
       arXiv: 2203.00236 [cs, eess].
[31]   V. A. Traag, L. Waltman, and N. J. van Eck. “From Louvain to Leiden: Guaranteeing Well-
       Connected Communities”. In: Scientific Reports 9.1 (2019), p. 5233. doi: 10.1038/s41598-0
       19-41695-z.
[32]   K. Vishnubhotla, A. Hammond, G. Hirst, and S. M. Mohammad. The Emotion Dynamics
       of Literary Novels. 2024. arXiv: 2403.02474 [cs].
[33]   G. B. Wilson. “Levels of Achievement in Acting”. In: Educational Theatre Journal 3.3
       (1951), pp. 230–236. doi: 10.2307/3204063. JSTOR: 3204063.
[34]   Y. Yu, Y. Hao, and P. Dhillon. “Unpacking Gender Stereotypes in Film Dialogue”. In: Social
       Informatics. Ed. by F. Hopfgartner, K. Jaidka, P. Mayr, J. Jose, and J. Breitsohl. Cham:
       Springer International Publishing, 2022, pp. 398–405. doi: 10.1007/978- 3- 031- 19097- 1
       \_26.
[35]   S. M. Zahiri and J. D. Choi. Emotion Detection on TV Show Transcripts with Sequence-based
       Convolutional Neural Networks. 2017. doi: 10.48550/arXiv.1708.04299. arXiv: 1708.04299
       [cs].
[36]   J. Zhao, T. Zhang, J. Hu, Y. Liu, Q. Jin, X. Wang, and H. Li. “M3ED: Multi-modal Multi-
       scene Multi-label Emotional Dialogue Database”. In: Proceedings of the 60th Annual Meet-
       ing of the Association for Computational Linguistics (Volume 1: Long Papers). Ed. by S.
       Muresan, P. Nakov, and A. Villavicencio. Dublin, Ireland: Association for Computational
       Linguistics, 2022, pp. 5699–5710. doi: 10.18653/v1/2022.acl-long.391.
[37]   F. Zhou and F. Pianzola. “Evaluation and Alignment of Movie Events Extracted via Ma-
       chine Learning from a Narratological Perspective”. In: 2023 Computational Humanities
       Research Conference, CHR 2023. CEUR Workshop Proceedings (CEUR-WS. org). 2023,
       pp. 49–62.




                                              199