<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Is Cinema Becoming Less and Less Innovative With to measure cultural innovation Time? Using neural network text embedding model ⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>EdgarDubourg</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>AndreiMogoutov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>NicolasBaumard</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institut Jean Nicod, Département d'études cognitives, Ecole normale supérieure, Université PSL, EHESS</institution>
          ,
          <addr-line>CNRS, 75005 Paris</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <fpage>676</fpage>
      <lpage>686</lpage>
      <abstract>
        <p>Current discourse re昀氀ects a growing skepticism towards contemporary popular culture, speci昀椀cally the realm of cinema, with an emerging consensus that its creative capacity is on a waning trajectory. This study introduces a novel approach which employs natural language processing techniques and embedding methods to measure semantic novelty of cultural items' descriptions. We apply this methodology to cinema, analyzing plot summaries of over 19,000 movies from the United-States spanning more than a century. Our measure's robustness is validated through a series of tests, including a 昀椀t with a genrebased novelty score, a manual inspection of 昀椀lms identi昀椀ed as highly innovative, and correlations with award recognitions. The application of our Innovation Score reveals a compelling pattern: an increase in the rate of cinematic innovation throughout the 20th century, followed by a stabilization in the rate of innovation in the 21st, despite an ever-growing production of 昀椀lms. Contrary to the o昀琀en-voiced lament that cinema is losing its innovative edge, our study suggests that the level of innovativeness in cinema is not in decline.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;innovation</kwd>
        <kwd>creativity</kwd>
        <kwd>culture</kwd>
        <kwd>text embedding</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The 昀椀lm industry has been the subject of numerous debates regarding its perceived decline
in innovation. Critics and audiences alike have voiced concerns about the increasing
prevalence of sequels, remakes, and franchise 昀椀lms, arguing that these trends would re昀氀ect a lack
of originality and creativity. These apprehensions have been ampli昀椀ed by the publicized
sentiments of esteemed 昀椀lmmakers such as Francis Ford Coppola, who likened popular 昀椀ctions
to “prototypes made over and over and over again,” Ken Loach, who compared them to
“commodities like hamburgers,” or Alejandro Iñárritu, who dismissed them as “basic and simple.”.
The apprehensions over the diminishing creative vigor in cinema are not con昀椀ned to academic
or elite circles; they resonate deeply with the broader public. A quick glance at social media,
椀昀lm forums, and audience reviews reveals a torrent of sentiments expressing disappointment
with the perceived stagnation of storytelling, reliance on formulaic plotlines, and the
increasing tendency to prioritize pro昀椀t over artistic innovation. But is it the case? Is innovation in
cinema on decline?</p>
      <p>
        The challenge in de昀椀ning innovation, and in determining whether a product is innovative,
lies in its subjective nature. We de昀椀ne innovation as what is novel for humanity at large, as
opposed to novelty, which is what is contextually novel for an individu1a8l,1[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This
distinction is crucial because it shi昀琀s the focus from individual perceptions to a collective level. Prior
research has attempted to quantify innovation in 昀椀lm by analyzing the unique combinations
of IMDb genres [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] or IMDb plot keywords1[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. However, these methods have inherent
limitations in tracking innovation over time. They heavily rely on metadata that tends to be more
abundant and precise for recent movies. This can result in a skewed perspective, as older 昀椀lms
o昀琀en lack comprehensive metadata. Other measures that aimed at quantifying innovation in
other creative domains such as the arts, technology, or science, relied on creaitnivdeividuals
(e.g., their interaction, see 1[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]; their number and productivity, see2[]; their place of birth and
movement, see [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]; their reputation, see 4[]).
      </p>
      <p>In this paper, we develop and apply a computational measure of innovation to movies based
on summaries, which are standardized and rather homogenous in both IMDb and Wikipedia.
This new measure is straightforwardly applicable, not to individuals, but to cultural products
and aims at measuring theirobjective level of innovativeness. It could in principle be applied in
di昀erent cultural domains, in di昀erent periods and countries, on di昀erent human productions
such as scienti昀椀c papers, patented technologies, or literary novels—or any other products with
textual descriptive metadata.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology: The computation of the cultural innovation score</title>
      <p>
        The Sentence-BERT (SBERT) algorithm 1[
        <xref ref-type="bibr" rid="ref20 ref3 ref3 ref7">3, 3, 7, 20</xref>
        ] is a robust tool for natural language
processing, widely utilized in applications ranging from text classi昀椀cation to information retrieval.
SBERT is built upon pre-trained transformer models like BERT and RoBER3T,a7][, which have
been trained on vast text datasets and have found applications in literary text comprehension
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Transformer models, with their self-attention mechanism, are adept at weighting the
significance of di昀erent parts of input data, making them highly e昀ective for tasks such as language
translation, text summarization, and question-answerin1g2[
        <xref ref-type="bibr" rid="ref14">, 14</xref>
        ].
      </p>
      <p>SBERT excels at computing semantic proximity between words, phrases, or even entire
paragraphs. It learns the contextual relationships between words (i.e., word embeddi9n,g1;0[]) and
can calculate the semantic similarity between new and existing text based on shared context.
This allows SBERT to accurately determine which words, sentences, or paragraphs are most
similar to each other, even if they are not exact matches. For instance, in the context of
cinema, SBERT could compute the semantic distance between the plot summaries oStfar Wars IV
(1977) and Star Trek: The Motion Picture (1979), recognizing their relatedness and scoring them
accordingly.</p>
      <p>In our measure of cultural innovation, we utilize SBERT to encode descriptions of cultural
products into 昀椀xed-length vectors that encapsulate the semantic meaning of the descriptive
text. These vectors are then used to compute the pairwise cosine similarity between them,
serving as a measure of their semantic similarity. The advantage of using SBERT over
traditional bag-of-words models is its ability to capture the meaning of the text, rather than just
the frequency of words, which is particularly useful when dealing with short texts and texts
coming from di昀erent periods or di昀erent contexts—where words di昀er although meaning is
similar.</p>
      <p>To compute our measure of cultural innovation for each individual cultural product, we
椀昀rst encode their description (here, movie plots) into high-dimensional vectors using SBERT.
We then compute the cosine similarity between each vector and all previous vectors. That
is, we compute the similarity between each product and all previous products. By reversing
this score, we transform the resulting similarity values into distance scores. Therefore, the
椀昀nal score, for each product, is computed asthe average of its distance scores from all previous
products. Building on this methodology, it is important to note that our measure inherently
captures the increasing di昀케culty of innovation as more products are released within the same
domain. As the number of preceding products grows, the space for unique, unexplored ideas
naturally shrinks, making it increasingly challenging to create something truly innovative.</p>
      <p>In the domain of movies, here, the innovation score for a given movie thus quanti昀椀es how
di昀erent its summary is from all previous movies in the dataset. This approach allows us to
capture the level of novelty of a movie plot compared to the plot of all movies having been
previously released.</p>
      <p>
        Our measure of innovation is conceptually similar to other methods that compute innovation
by assessing semantic distances between individual cultural products and their predecessors
within the same domain. For example, in the realm of French theater, Ca昀椀ero and Gabay1][
have employed a similar approach, as have Kelly and colleagu6e]si[n the analysis of
technological patents (see 1[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], for an application).
      </p>
      <p>We can further formalize this measure. Given a set of cultural products, each with a
description   , the Innovation Score (IS) for th e-th product is calculated as:
   =
1</p>
      <p>∑(1 −
 − 1 &lt;</p>
      <p>⋅  
||  || ⋅ ||  ||
)
where   = SBERT(  ) is the high-dimensional vector representation of the descriptio n
obtained using the Siamese-BERT (SBERT) algorithm⋅, denotes the dot product, and|| ||
denotes the Euclidean norm of vector . IS essentially measures how distinct or innovative a
cultural product’s description is compared to the descriptions of all the previous products in
the set. A higher IS would indicate that the product’s description is more unique and
innovative within the given set. Conversely, a lower IS would suggest that the product’s description
shares more similarities with the descriptions of previously seen products.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Validity Check: Evaluation of our Measure of Innovation</title>
      <p>Our study utilizes a comprehensive dataset compiled from IMDb and Wikipedia, encompassing
metadata for 19,254 movies produced in the United States. The IMDb data, obtained directly</p>
      <sec id="sec-3-1">
        <title>3.1. Robustness Across Di昀erent Sources of Description</title>
        <p>We turned to IMDb plot summaries, which di昀er markedly from their Wikipedia counterparts
in length and standardization. Despite these di昀erences, our Innovation Score remained
consistent, demonstrating a signi昀椀cant positive correlation between the scores derived from both
sources ( = .28 ,  &lt; .001 ). This result underscores the adaptability of our measure, capable of
capturing innovation irrespective of the descriptive metadata available.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Robustness Across Di昀erent Random Seeds</title>
        <p>To further assess the robustness of our Innovation Score, we altered the random seed used
in the calculation process. This analysis consistently revealed a strong positive correlation
( = 0.89 ,  &lt; .001 ) between the Innovation Scores obtained using two di昀erent random seeds.
This substantial correlation reinforces the reliability and stability of our Innovation Scores,
demonstrating its consistency even when varying the initial randomization.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Robustness Across Di昀erent Timeframes</title>
        <p>The notion of cultural forgetting would suggest that a movie can appear innovative even if its
narrative resembles older 昀椀lms, if it is at least di昀erent fromrecent ones. To examine this, we
calculated three new Innovation Scores, progressively considering a narrower temporal
window: we computed the average distance of each movie from movies released 10 years before,
5 years before, and just 1 year before. Strikingly, our analysis revealed that the level of
innovation remained nearly identical across these di昀erent timescales (for all correlat io&gt;ns.9,8 ,
 &lt; .001 ). This suggests that a movie’s innovativeness is a consistent trait, irrespective of the
speci昀椀c timeframe under scrutiny.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Qualitative Observation</title>
        <p>Upon manual inspection, we found that our algorithm indeed identi昀椀ed movies widely
acclaimed as innovative, such as2001: A Space Odyssey, Pulp Fiction, and Interstellar, as highly
innovative (see Figure 2.A.). However, it is important to note that these are cherry-picked
examples, and we could have chosen others that would not have aligned with our intuition.
For instance, whileInception is also widely considered innovative, it received a relatively low
Innovation Score in our analysis. This qualitative examination serves as an initial validation
to check that some scores 昀椀t our intuitions.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Correlation with Another Measure of Innovation</title>
        <p>To ensure the robustness of our measure beyond qualitative inspection, we compare our
Innovation Score with a Novelty Score derived from a method proposed by Luan and K8im], [
which gauges the uniqueness of a movie’s genre combination relative to preceding 昀椀lms. This
Novelty Score is genre-based: it rewards 昀椀lms introducing rare combinations of genres. We
used this Novelty Score as a benchmark to evaluate the e昀ectiveness of our Innovation Score in
capturing a movie’s deviation from genre conventions. As anticipated, we found a signi昀椀cant
positive correlation between our Innovation Score and the genre-based Novelty Sco=re.0(4 ,
 &lt; .001 ), bolstering the external validity of our measure (see Figure 2.B.).</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.6. Correlation with Movie Genres</title>
        <p>We conducted multiple two-sample t-tests comparing the aggregated Innovation Scores of
movies within a speci昀椀c genre to those outside of it (with Bonferroni correction for multiple
testing). This test allowed us to discern whether there was a signi昀椀cant di昀erence in the mean
level of innovation between movies belonging to a genre and all the other ones (see Figure 2.C.).
Genres that encompass formulaic narrative plots, such as Film Noir, Mystery, Crime, Sport, and
Thriller, exhibit lower average Innovation Scores. This phenomenon arguably occurs because,
to belong to these genres, a given movie needs to adhere to speci昀椀c narrative conventions. In
contrast, genres like Adventure, Action, and History, characterized by non-speci昀椀c themes,
allow for greater innovation because they accommodate a broader spectrum of narratives that
can deviate from traditional storytelling structures. While it may seem counterintuitive at 昀椀rst,
the high average Innovation Scores in War 昀椀lms can be attributed to their ability to draw from
various historical events and kinds of warfare. Science Fiction’s high average Innovation Scores
can be attributed to its futuristic focus and the inherent audience expectation for novelty in
Science Fiction movies.</p>
      </sec>
      <sec id="sec-3-7">
        <title>3.7. Correlation with Awarded Movies</title>
        <p>Awarded movies have higher average Innovation Scores. Building on the common intuition
that award juries, who are cinema experts, tend to reward innovation in cinema, we sought to
investigate the relationship between our Innovation Scores and movie awards. We created a
binary variable indicating whether a movie had won at least one award, based on mentions in
Wikipedia pages, and conducted a two-sample t-test to compare the innovation levels of
awardwinning movies and those without awards. This analysis was extended to speci昀椀c awards like
the Academy Awards, Sundance Awards, and the Palme d’Or (with Bonferroni correction for
multiple testing). Our 昀椀ndings revealed that, in general, award-winning movies tend to have
higher Innovation Scores (see Figure 2.D.). Notably, movies recognized at the Sundance Film
Festival, known for its focus on innovative independent 昀椀lms, were associated with higher
innovation levels. However, this correlation was not observed for other speci昀椀c awards, such
as the Independent Spirit Awards. This discrepancy could underlie the idea that our Innovation
Score captures a speci昀椀c type of narrative innovation, which may not align with the criteria
used by all award bodies. Nevertheless, the overall trend supports the validity of our Innovation
Score as a measure of innovation in cinema.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results: The Evolution of Innovation in Cinema</title>
      <p>In our exploration of the temporal dynamics of innovation in cinema, we aggregated the
Innovation Scores by year and conducted a series of regression analyses. We 昀椀tted three models:
a linear model, a quadratic model, and a logarithmic model, each with the mean Innovation
Score as the dependent variable and the year as the independent variable.</p>
      <p>The linear model, which assumes a constant rate of change in innovation over time,
accounted for approximately 18% of the variance in the data. The logarithmic model, which
posits a decelerating rate of innovation, performed similarly to the linear model, explaining
approximately 18% of the variance. However, the quadratic model, which allows for a
changing rate of innovation, performed signi昀椀cantly better, explaining about 28% of the variance.
To further compare these models, we calculated the Akaike Information Criterion (AIC) and
the Bayesian Information Criterion (BIC), both of which balance the goodness-of-昀椀t of a model
with its complexity. Lower values of AIC and BIC indicate a better model. The quadratic model
outperformed the other two models on both criteria, further supporting its superiority (Figure
3.A.).</p>
      <p>The quadratic model suggests a non-linear relationship between time and innovation in
cinema. The positive linear term in the model indicates an overall increase in innovation over
the years, while the negative quadratic term (lower in magnitude) suggests a slowing down of
this increase (Figure 3.B.). Speci昀椀cally, by observing the plot of the 昀椀tted quadratic model, we
can infer that innovation in cinema experienced a surge throughout the 20th century.
However, this rate of increase appears to have decelerated and reached a plateau in recent years,
indicating a stabilization of innovation levels in the cinematic landscape.</p>
      <p>In addition to our regression analyses, we wanted to explore to what extent the production
of 昀椀lms over the course of history deviates from what it would have been if the 昀椀lms had
appeared randomly over time. We used a Monte Carlo simulation approach. This simulation
involved generating 1000 datasets, each with movies randomly shu昀툀ed across di昀erent years
while keeping the number of movies per year constant. The results of this Monte Carlo
simulation are striking. They reveal a decrease in the average Innovation Score during the initial
years, followed by a relatively constant, lower level of innovation per year. The reason is
straightforward: the more 昀椀lms already exist, the more di昀케cult it is to innovate on a purely
random basis. This pattern contrasts sharply with the actual data, which showed an increase
in innovation throughout the last century.</p>
      <p>This suggests that if movies were randomly distributed in time, without consideration of
what came before, they would produce movies that share similarities with previous eras purely
by chance, leading to a 昀氀atter and lower average Innovation Score. In contrast, in the real
dataset, they seem to actively strive to innovate and di昀erentiate their work from what has
come before, leading to an overall increase in innovation over time. This observation, therefore,
highlights the strong connection between movie production, creativity, and the in昀氀uence of
past cinematic trends. Filmmakers draw from the past while attempting to break away from
prevailing storylines, resulting in the observed increase in innovation in the real dataset as
opposed to the simulated ones.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Our analysis of over 19,000 movies spanning more than a century has yielded fascinating
insights into the trajectory of cinematic innovation. We observed a signi昀椀cant increase in
innovation throughout the 20th century, underscoring the era’s reputation as a period of rapid
creative evolution. Thus, contrary to the o昀琀en-voiced lament that cinema is losing its
innovative edge, our study suggests that the level of innovativeness in cinema is not in decline. In fact,
according to our model, the level of innovation today is as high as it was during the golden era
of cinema in the 1950s. This implies that the use of formulaic plots is not more prevalent now
than it was in the past.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ca昀椀ero</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Gabay</surname>
          </string-name>
          . “
          <article-title>Rise and Fall of Theatrical Genres in Early Modern France: a Centroid-Based Approach”</article-title>
          . InP:ublisher: arXiv (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B. d.</given-names>
            <surname>Courson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Thouzeau</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Baumard</surname>
          </string-name>
          . “
          <article-title>Quantifying the scienti昀椀c revolution”</article-title>
          .
          <source>In: Evolutionary Human Sciences</source>
          <volume>5</volume>
          (
          <year>2023</year>
          ),
          <year>e19</year>
          . doi:
          <volume>10</volume>
          .1017/ehs.
          <year>2023</year>
          .
          <article-title>6</article-title>
          . url: https://www.c ambridge.org/core/journals/evolutionary
          <article-title>-human-sciences/article/quantifying-the-scie ntific-revolution/60249C6B9DF636D2EC8446F6B7E454F.8</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          . “BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding”</article-title>
          . In: (
          <year>2018</year>
          ).
          <year>do10i</year>
          :.48550/arxiv.
          <year>1810</year>
          .
          <volume>0</volume>
          4805. url: https://arxiv.org/abs/
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S. P.</given-names>
            <surname>Fraiberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sinatra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Resch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Riedl</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.-L.</given-names>
            <surname>Barabási</surname>
          </string-name>
          . “
          <article-title>Quantifying reputation and success in art”</article-title>
          .
          <source>In:Science 362.6416</source>
          (
          <year>2018</year>
          ), pp.
          <fpage>825</fpage>
          -
          <lpage>829</lpage>
          . doi:
          <volume>10</volume>
          .1126/science.aa u7224. url: https://www.sciencemag.org/lookup/doi/10.1126/science.aau72
          <volume>2</volume>
          .
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Breithaupt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kübler</surname>
          </string-name>
          , and T. T. Hills. “
          <article-title>Quantifying the retention of emotions across story retellings”</article-title>
          .
          <source>InS:cienti昀椀c Reports 13.1</source>
          (
          <issue>2023</issue>
          ), p.
          <fpage>2448</fpage>
          . doi:
          <volume>10</volume>
          .1038/s41598-02
          <fpage>3</fpage>
          -
          <lpage>29178</lpage>
          -8. url: https://www.nature.com/articles/s41598-023-
          <fpage>29178</fpage>
          -.
          <fpage>8</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Kelly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Papanikolaou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Seru</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Taddy</surname>
          </string-name>
          . “
          <article-title>Measuring Technological Innovation over the Long Run”</article-title>
          . In:American Economic Review: Insights (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          . “
          <article-title>RoBERTa: A Robustly Optimized BERT Pretraining Approach”</article-title>
          . In: arXiv:
          <year>1907</year>
          .11692 [cs] (
          <year>2019</year>
          ). eprint:
          <year>1907</year>
          .11692. url: http://arxiv.org/abs/
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luan</surname>
          </string-name>
          and
          <string-name>
            <surname>Y. J. Kim. “</surname>
          </string-name>
          <article-title>An integrative model of new product evaluation: A systematic investigation of perceived novelty and product evaluation in the movie industry”</article-title>
          .
          <source>In: PloS One 17.3</source>
          (
          <issue>2022</issue>
          ),
          <year>e0265193</year>
          .
          <source>doi: 10.1371/journal.pone.026519.3</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          . “
          <article-title>E昀케cient Estimation of Word Representations in Vector Space”</article-title>
          . In:Publisher: arXiv Version Number:
          <volume>3</volume>
          (
          <year>2013</year>
          ). doi:
          <volume>10</volume>
          .48550/arxiv .1301.3781. url: https://arxiv.org/abs/1301.3781.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Manning</surname>
          </string-name>
          . “Glove:
          <article-title>Global Vectors for Word Representation”</article-title>
          .
          <source>In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          .
          <source>Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          . Doha, Qatar: Association for Computational Linguistics,
          <year>2014</year>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          . doi:
          <volume>10</volume>
          .3115/v1/
          <fpage>D14</fpage>
          -1162. url: http://aclweb.org/anthology /D14-1162.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Posch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schulz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Henrich</surname>
          </string-name>
          . “Surname Diversity,
          <article-title>Social Ties and Innovation”</article-title>
          .
          <source>In: SSRN Electronic Journal</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sachan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Felix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Padmanabhan</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Neubig.</surname>
          </string-name>
          “
          <article-title>When and Why Are PreTrained Word Embeddings Useful for Neural Machine Translation?”PIrnoc:eedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          , Volume
          <volume>2</volume>
          (
          <string-name>
            <given-names>Short</given-names>
            <surname>Papers</surname>
          </string-name>
          <article-title>)</article-title>
          .
          <source>Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>2</volume>
          (
          <string-name>
            <surname>Short</surname>
            <given-names>Papers). New</given-names>
          </string-name>
          <string-name>
            <surname>Orleans</surname>
          </string-name>
          , Louisiana: Association for Computational Linguistics,
          <year>2018</year>
          , pp.
          <fpage>529</fpage>
          -
          <lpage>535</lpage>
          .
          <year>do1i0</year>
          :.
          <year>1865</year>
          3/v1/
          <fpage>N18</fpage>
          -2084. url: http://aclweb.org/anthology/N18-208.
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          and
          <string-name>
            <surname>I. Gurevych.</surname>
          </string-name>
          “
          <article-title>Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation”</article-title>
          .
          <source>InP:roceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          .
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          .
          <source>Online: Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>4512</fpage>
          -
          <lpage>4525</lpage>
          .
          <year>doi1</year>
          :
          <fpage>0</fpage>
          .18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>365</volume>
          . url: https://www.aclweb.org/anthology/2020.emnlp-main.
          <volume>36</volume>
          .5
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          and
          <string-name>
            <given-names>I.</given-names>
            <surname>Gurevych.</surname>
          </string-name>
          Sentence-BERT:
          <article-title>Sentence Embeddings using Siamese BERTNetworks</article-title>
          .
          <year>2019</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1908</year>
          .
          <volume>10084</volume>
          . arXiv:
          <year>1908</year>
          .10084[cs]. url: http://arxiv .org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Ahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mirsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-L.</given-names>
            <surname>Barabási</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Helbing</surname>
          </string-name>
          . “
          <article-title>Quantitative social science. A network framework of cultural history”S.cInie:nce</article-title>
          (New York, N.Y.)
          <volume>345</volume>
          .6196 (
          <year>2014</year>
          ), pp.
          <fpage>558</fpage>
          -
          <lpage>562</lpage>
          . doi:
          <volume>10</volume>
          .1126/science.1240064.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>M. Sera昀椀nelli</surname>
          </string-name>
          and G. Tabellini. “
          <article-title>Creativity over time and space”</article-title>
          .
          <source>InJo:urnal of Economic Growth 27.1</source>
          (
          <issue>2022</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>43</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10887-021-09199-6. url: https://doi.org/10.1 007/s10887-021-09199-6.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sreenivasan</surname>
          </string-name>
          . “
          <article-title>Quantitative analysis of the evolution of novelty in cinema through crowdsourced keywords”</article-title>
          .
          <source>InS:cienti昀椀c Reports 3.1</source>
          (
          <issue>2013</issue>
          ), p.
          <fpage>2758</fpage>
          . doi:
          <volume>10</volume>
          .1038/srep0275 8. url: http://www.nature.com/articles/srep0275.8
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Tacchella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Napoletano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Pietronero</surname>
          </string-name>
          . “
          <article-title>The Language of Innovation”</article-title>
          .
          <source>PInlo:s One 15.4</source>
          (
          <issue>2020</issue>
          ),
          <year>e0230107</year>
          . doi:
          <volume>10</volume>
          .1371/journal.pone.
          <volume>023010</volume>
          .7url: https://journals.plos .org/plosone/article?id=
          <volume>10</volume>
          .1371/journal.pone.
          <volume>0230</volume>
          .107
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>F.</given-names>
            <surname>Tria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Loreto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. D. P.</given-names>
            <surname>Servedio</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Strogatz</surname>
          </string-name>
          . “
          <article-title>The dynamics of correlated novelties”</article-title>
          .
          <source>In:Scienti昀椀c Reports 4.1</source>
          (
          <issue>2014</issue>
          ), p.
          <fpage>5890</fpage>
          . doi:
          <volume>10</volume>
          .1038/srep05890. url: https://w ww.
          <source>nature.com/articles/srep0589.0</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Polosukhin.</surname>
          </string-name>
          “Attention Is All You Need”.
          <source>InP:ublisher: arXiv Version Number: 5</source>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .48550/arxiv.1706.03762. url: https://arxiv.org/abs/1706.03762.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>