<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Automated Video Game Commentary Using Generative AI</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Noah Renella</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Markus Eger</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cal Poly Pomona</institution>
          ,
          <addr-line>Pomona</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In recent years, video game streaming has enjoyed ever-growing popularity. While streaming gameplay, a streamer is also expected to provide commentary on gameplay and otherwise entertain the audience. The mental burden for providing such commentary while also playing the game may provide a barrier of entry for new streamers. In this paper, we present an approach to automatically generate streaming commentary during gameplay for the game League of Legends. Our system recognizes key events, and uses generative AI services to generate voice output. We also present a preliminary evaluation of our system, both in quantitative terms of event recognition accuracy, as well as qualitatively by discussing the generated commentary.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        work that recognizes in-game events, text generation via
API-access to ChatGPT, and finally audio-generation
usVideo game streaming has become one of the most pop- ing the FakeYou API. We increase the engagement and
ular forms of online entertainment [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. It pairs a host, variety of the generated commentary by having it be
or “streamer”, who plays a game and provides commen- delivered in the style and voice of (fictional) characters.
tary with an audience that watches the streamer’s game- Our system can also provide color commentary to keep
play, listens to the commentary, and interacts with the the audience engaged in periods of low in-game
activstreamer in a chat. For the streamer, this split attention ity. Our approach is meant to augment a streamer’s own
between trying to perform well in the game (especially in commentary, relieving them of some cognitive load, and
a competitive game setting), explaining the game to the allowing them on focusing on interacting with the
auaudience, and interacting with it, can be quite daunting. dience, which our system does not currently cover. We
At the same time, there are many events that occur reg- provide both quantitative results measuring how well the
ularly in a given game, making the task of the streamer system can recognize the relevant game events, as well
repetitive and perhaps less engaging. as qualitative results of output generated by the system.
      </p>
      <p>
        If viewed as a spectator sport, providing commentary
during streaming is similar to the commentary provided
by professional sports casters during e.g. a football match 2. Background and Related Work
(American or not). This commentary typically consists
of two components: An objective play-by-play reporting, Providing commentary for gameplay is a challenging task
as well as additional color commentary to provide back- that has found interest for a variety of diferent
applicaground information, statistics or simply entertainment tions. Shah et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] describe a system that can generate
during lulls in gameplay. In professional sports casting, commentary for Minecraft. The authors use a data set
these two components are often filled by two separate of three videos of Minecraft game-play with comments
individuals, and each follows its own speech patterns to learn which frames should go with which type of
and rules [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Video games, in the form of eSports tour- commentary using a combination of clustering and
Connaments, follow similar commentary patterns. In online volutional Neural Networks. Similar in data collection,
streaming, however, there is often only one streamer that but not in the overall approach, Li et al. ([
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]) collected
has to take on both roles, providing insights into the existing video commentary for the game Getting over it
running game as well as entertaining their audience. with Bennet Foddy and use a sequence-to-sequence model
      </p>
      <p>
        In this paper, we present a preliminary approach to au- to learn how to generate commentary. Our approach
diftomatically provide audio commentary during gameplay fers in two significant ways: First, we introduce existing
of the game League of Legends. Our approach consists domain knowledge into the system, to identify which
of a pipeline of components consisting of a Neural Net- events warrant commentary. Second, instead of relying
on existing commentary videos only, we employ a
general generative text model, ChatGPT, by constructing the
OAcIItDobEerW0o8r,k2s0h2o3p,UonnivEexrpsietryi mofeUnttaalh,AUrttiafichi a,lUISnAtelligence in Games, appropriate prompts.
$ ntrenella@cpp.edu (N. Renella); meger@cpp.edu (M. Eger) Commentary generation has been of interest for
en© 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License tertainment purposes outside video games as well. For
CPWrEooUrckReshdoinpgs IhStpN:/c1e6u1r3-w-0s.o7r3g ACttEribUutRion W4.0oInrtekrnsahtioonpal (PCCroBYce4.0e).dings (CEUR-WS.org)
the RoboCup robotic soccer competition, for example, provide diferent commentary. An example of such event
commentary generation was seen as a grand challenge can be seen in figure 2. Commentary a viewer might
as early as 1998 [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]. For this particular domain, the expect for this event could be “We are going to lose.” or
key challenge addressed by the various approaches was “We have to play better if we want to win”.
mainly identifying key game moments, while the
commentary generation was based on predefined templates. 2.2. Generative AI
While RoboCup, and in particular the simulation league,
constitutes a middle-point between real life sports and
video games, other authors have also worked on
generating commentary for Baseball [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The focus here,
too, was on being able to even identify key moments in
gameplay. In our work, event detection plays a relevant
role as well, but the specifics difer due to our application
domain, which we will briefly discuss in the next section.
      </p>
      <p>
        To generate the commentary we utilize ChatGPT, an
advanced language model developed by OpenAI1. The
model is based on a transformer-based architecture
[
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ] and is trained on a large corpus of text data which
allows it to generate a wide range of plausible-looking,
natural text. ChatGPT is capable of generating relevant
and suitable responses to prompts a user provides which
allows it to engage in human-like interactions with the
2.1. League of Legends user. Its versatility extends to various other applications,
such as chatbots, content generation [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], and many other
The game we are targeting for commentary generation is use cases.
      </p>
      <p>
        League of Legends . League of Legends is a popular multi- To convey the generated commentary to the audience,
player online battle arena game developed and published we use FakeYou2, an open-source text-to-speech reader
by Riot Games [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The game matches two teams of five equipped with voice cloning capabilities. This tool
utiplayers each, where each player controls a champion- lizes machine learning technology to process and
replicharacter, and each team works together to destroy the cate voices, allowing us to incorporate these voices into
opposing team’s base. Each base is defended by a se- voice-overs. By employing artificial intelligence and deep
ries of turrets, as well as computer-controlled characters. learning techniques, FakeYou comprehends and analyzes
Players can attack each other, the turrets, or the base the speech pattern as well as the inflection and
intonaitself, using diferent attacks and abilities unique to each tion, for a specific character. As a result, it can generate
champion. There are many events that are of interest for remarkably authentic audio output that resembles the
commentary generation that take place during a game, desired character. The FakeYou API provides us with over
such as an ally being slain, a player achieving three con- 3000 voices of popular people and characters, which can
secutive kills of enemy players (called a “triple kill”), and, be used to read prompts and deliver high-quality audio
ultimately, winning or losing the game. An example of output.
how such an event is represented in the game can be seen
in figure 1.
      </p>
      <p>While watching a stream of the game, a viewer seeing 3. Approach
the event in figure 1 may expect positive commentary In order to generate commentary, we use three separate
such as “Well played team!” or “We are one step closer components: Event detection, commentary generation,
to victory!”. In other cases, the events that take place
are not in favor of the player which is where we need to
1https://chat.openai.com/
2https://fakeyou.com/
and voice realization. Before we go into detail about
how each of these components work, we will briefly
describe the overall structure of our system. Our system
attempts to balance between providing neither too little,
nor too much commentary. As there may be periods of
low activity in the game, the system will generate color
commentary to fill these gaps. On the other hand, if there
are many diferent events that occur during gameplay,
our system will prioritize more important events. The
general outline for our system is shown in listing 1.</p>
      <p>WHILE streamed v i d e o i s p l a y i n g :
c u r r e n t _ f r a m e = 1 s t frame o f th e 30 per second
c u r r e n t _ e v e n t = d e t e c t _ e v e n t ( c u r r e n t _ f r a m e )
IF c u r r e n t _ e v e n t has an e v e n t :</p>
      <p>FOR 5 seconds :</p>
      <p>IF h i g h e r p r i o r i t y e v e n t o c c u r s :</p>
      <p>r e p l a c e c u r r e n t _ e v e n t
IF c u r r e n t _ e v e n t has an e v e n t and 5 s econds</p>
      <p>have p a s s e d :
commentary = g e n e r a t e ( c u r r e n t _ e v e n t )
w a v _ f i l e = t e x t −to − speech ( commentary )
p l a y ( w a v _ f i l e )
IF 25 seco nds p a s s without an e v e n t :
commentary = color_commentary ( )
w a v _ f i l e = t e x t −to − speech ( commentary )
p l a y ( w a v _ f i l e )</p>
      <p>Listing 1: System Psuedocode</p>
      <sec id="sec-1-1">
        <title>With this system, the audience will hear at least one</title>
        <p>generated commentary every 25-30 seconds. The system
follows a sequential process to provide event-based
commentary during a video stream. If an event is detected, a
5-second timer is initiated to observe if any other event
takes place that may take priority.</p>
        <p>Once the 5 second interval has passed, if no
higherpriority event occurs, the system generates commentary
related to the event using ChatGPT. The commentary
thus generated is then sent to the FakeYou API to
produce audio output for the chosen character, which is then
played. The time frame of 5 seconds for a higher priority
event was determined through some experimentation, to
give a trade-of between how to keep generated
commentary current, while still giving the game opportunity to
result in a higher-priority event. For example, the player
killing an opponent results in an event, which the system
ought to comment on. However, if the player
immediately kills a second enemy player, the resulting double
kill would be preferable to comment on. We will describe
the individual events below, as well as the priorities we
assigned them.</p>
        <p>If no event has occurred in a 25 second-window, the
system will generate color commentary instead. Unlike
event-based commentary which provides information
about what is occurring currently, color commentary
offers responses to audience messages or gives background
details or facts to fill the silence during a stream.
(3a): b_triple event header
(3a): section of b_triple event being annotated</p>
        <sec id="sec-1-1-1">
          <title>3.1. Model Training</title>
          <p>In order to be able to detect events in the game, we
handannotated a data set of frames with their corresponding
events, and then trained a Neural Network-based model
on this data set. The data set was created using 320 event
frames gathered from over 80 gameplay videos. These 320
frames are annotated using LabelImg,3 an open-source
tool used for annotating bounding boxes over regions
of interest in images used for object detection. Event
information shown on the screen consists of the event
type, team, as well as a representation of the champion or
champions involved in the event. Champions difer from
game to game so only the event text and art are annotated.
Figure 3 shows an example of an event notification in the
game.</p>
          <p>While the game provides a great number of events,
some of them are exceedingly rare to the point that they
did not show up in our collected data set in suficient
number. Instead, we opted to label only the most frequent
events for model training. We gave each event a label,
starting with a b or r to indicate which team the event
is happening for. b indicates the blue or ally team, while
r is the red, or enemy, team. For example r_triple
indicates that the enemy team got a triple kill.</p>
          <p>
            We use EficientDet [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ] as our object detector. This
architecture incorporates EficientNet [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ] as the backbone
network and integrates a BiFPN feature network. The
model is trained using TensorFlow Lite and the model
maker library.
          </p>
          <p>We trained two diferent models for comparison
purposes: The first model is trained on 224 labelled images of
the 14 most common events, while the second model was
trained on the full 320 labelled images of all 20 even1ts4
present in our data set. As some events happen very
rarely, we used these two diferent models to evaluate
the diference. Both models were trained with a batch
size of 4 for 300 epochs.</p>
        </sec>
      </sec>
      <sec id="sec-1-2">
        <title>Once an event is detected, the system waits for 5</title>
        <p>seconds before generating commentary. This allows for
the game to play out for any other more important events
3.2. Event Recognition to take place. If a new event does occur and is detected,
its priority will be compared with the current event to
To actually detect events within a game of League of check if needs to be replaced. Once the 5 seconds have
Legends using the trained models, OpenCV is utilized to passed the system will take the highest priority event
read the video and present each frame to the audience. and generate the commentary using ChatGPT.
A standard video stream will display at least 30 frames
per second, and each frame could be utilized for event
detection. However, processing 30 or more frames per 3.3. Commentary Generation
second requires significant computational power and is ChatGPT operates by taking a prompt and generating
impractical. Since each event period lasts longer than a response based on it. For instance, when given the
one second, we only apply the object detector to one prompt “React to yourself getting a double kill in the
frame per second. If an event is detected, further checks game League of Legends”, it might respond with “I can
are performed to confirm if commentary generation is not believe I got a double kill, we are one step closer to
necessary. winning”. This dynamic approach enables us to generate</p>
        <p>To ensure that commentary is generated for the most realistic, variable commentary in real-time according to
significant event, we assigned priorities to each event, the most significant event. In order to provide an actual
depending on their impact on the game. Winning the prompt to ChatGPT, we assign a prompt template to each
game is one of the most important events to take place event, which can be seen in listing 3.
which is why we would want to prioritize this event over
any others. Some of the assigned priorities are shown in
listing 2.</p>
        <p>With the dynamic generation of ChatGPT we can also
provide a character name to change the style of the
commentary being generated. For example, in our
experiments, we are using the character Rick Sanchez from
the TV show “Rick and Morty” because his speech
patterns and mannerism are defined and recognizable, and
provide additional variety to the commentary.</p>
        <p>Generating commentary for events is important as it
gives the audience insight into how the game is
progressing but there can be long periods of time where no event
takes place. To fill this silence we also define prompts
to generate color commentary, if no event is detected
within a 25 second period. These prompts, shown in
listing 4, are random messages or events outside the game a
streamer may encounter while playing which does nicely
to fill in the silence.</p>
        <sec id="sec-1-2-1">
          <title>3.4. Speech Generation</title>
          <p>Finally, once an event is detected and commentary is
generated we convey the commentary to the audience
as audio by using FakeYou. The generated commentary
prompt is passed to the FakeYou API together with the
character token used to generate the commentary for.
The resulting audio file is then played to the audience.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>4. Results</title>
      <p>(5a): b_double event before detection
(5a): b_double event after detection
“WUBALUBADUBDUB” and general speech pattern, and
provides the audience with relevant and valuable insight
to the events taking place in the game adding to their
viewing experience.</p>
      <p>An example of generated color commentary is: “I’m
feeling extra confident that I’m gonna win this game of
League of Legends. I know my skills, I know the map,
and I know how to outplay my opponents.” Although the
commentary is not specifically linked to an event
occurring within the game, it enhances the viewing experience
by providing additional insights, analysis, or entertaining
remarks.</p>
      <p>In order to determine how well the system is able to
process events, we have performed a quantitative
evaluation of the event detection, as well as a preliminary
qualitative evaluation of the resulting commentary.</p>
      <p>The final system can processes a video feed of League 4.1. Quantitative Evaluation
of Legends, while detecting the events that take place
in real time. During each detected event, if commentary To evaluate the two event detection model’s performance,
generation is necessary, it will generate commentary and we evaluated them on 4 prerecorded videos of League of
convey it using FakeYou voices. An example of such Legends gameplay. Each video consists of a number of
event detection can be seen in figure 5. events, which were hand-annotated in order to evaluate</p>
      <p>Using the Rick Sanchez character token, the gener- the correctness of the models’ predictions.
ated the commentary for the double kill shown in fig- Table 1 shows the results of a 11:44 minute segment
ure 5 is: “WUBALUBADUBDUB! Double kill! That’s of League of Legends gameplay which takes place
durwhat I’m talking about! Let’s keep this up and take ing the middle stages of the game. During this segment
the win!”. This commentary uses the character-typical 20 events were detected and it generated 11 color
com</p>
      <p>Table 2 shows the results of a 9:40 minute segment of
gameplay. This video takes place during the early stage of
the game where events take place less frequently. There
are more silent periods during the early stage which is
where the system generates color commentary. This is
why despite the shorter run time than table 1, there are
3 more generated color commentary.</p>
      <p>The 14-label model demonstrated a 77.14% accuracy
in detecting the events that occurred within the four
videos. It successfully identified the participating teams
in the events with 100% accuracy, even in cases where
the overall event being detected was incorrect.</p>
      <p>Although the 20-label model is more robust due its
wider range of labels, its overall accuracy was only
61.11%, lower than the 14-label model. This increased
label coverage did allow it to detect events not possible
for the 14-label model but it still had trouble detecting
the events correctly.</p>
      <p>In some cases the model may detect an incorrect event
occurring such as a b_double event being detected as a
b_triple event. The commentary being generated for the
detected event is, “Morty, did you see that? That was one
sweet triple kill!”, which is incorrect. This might
potentially confuse the viewer and provide them in inaccurate
information regarding the events occurring in the game.</p>
      <sec id="sec-2-1">
        <title>4.2. Qualitative Evaluation</title>
        <p>To test the quality of the generated commentary we
generated it in the style of Rick Sanchez from the television
show Rick and Morty. Rick Sanchez’s speech
mannerisms and patterns are very recognizable, allowing us to
check the generated commentary for similarities. Table
3 showcases the commentary generated for the video
of table 1. As can be seen, the generated commentary
makes frequent use of Rick’s catchphrase “Wubba lubba
dub dub” and other iconic quips, and generally echoes
his character on the show.</p>
        <p>Additionally, each generated response is tailored to the
specific event that was recognized, ofering the viewer or
listener tuning into the stream meaningful and relevant
commentary.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>5. Ethical Considerations</title>
      <p>
        The use of generative AI techniques is not without
(justified) controversy [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Legal questions regarding the
appropriate use of the training data will necessarily have
to be answered in an appropriate court of law, but beyond
that our responsibility as researchers is two-fold: One
one hand, the types of applications we use these
emerging technologies for can help shape their future direction.
On the other hand, services like ChatGPT and FakeYou
exist regardless of whether we use them in research or
not. Only by using them, however, can we investigate
their risks, biases and limitations.
      </p>
      <p>We address this dual responsibility in two ways: First,
our particular application focuses purely on
entertainment, but we nevertheless found it prudent avoid using
the voices of real people. Instead, our approach uses
fictional characters to avoid any (accidental) appearance
of impersonation. We will note, however, that these
capabilities exist, and require further discussion to ensure
their responsible use, if one exists.</p>
      <p>Second, as can be discerned from the provided output,
there are still challenges with using such an approach:
Without legal guidance, it is impossible to determine if
the output would violate copyright law (although “Wubba
Lubba Dub Dub” as a phrase itself does not appear to be
copyrighted). We also currently rely on OpenAI’s own
safeguard methods on the output, which is perhaps not
perfect [16]. Additionally, even though we intentionally
used a fictional character, there are also open questions
surrounding the likeness of the voice actor. As stated
above, while we believe it is important to investigate the
use of these emergent technologies, we will err on the
side of caution, and eschew releasing the code publicly
at this point.</p>
    </sec>
    <sec id="sec-4">
      <title>6. Conclusion and Future Work</title>
      <p>This paper addresses some of the challenges faced by
streamers while live streaming video games and
delivering meaningful commentary. We present an approach
based on a combination of several machine learning
models, including technologies such as ChatGPT and FakeYou
voices, aimed at assisting streamers. Our approach
involves the training of an event detection model on a
dataset of gameplay recordings, and using it to detect
said events in order to generate commentary using
ChatGPT and conveying the commentary using FakeYou. The
model created in our work was designed for the game
League of Legends, but our approach is applicable to any
game with specific events. We also present an evaluation
of our approach, quantitatively in terms of how accurate
the event detection is, as well as a qualitative evaluation
of the generated commentary.</p>
      <p>While our work presents encouraging initial results,
there is significant room for future improvement. One
key challenge we encountered is data collection, and
the necessity to hand-annotate gameplay footage. Some
events, such as penta-kills, are exceedingly rare, and
were not present in any of our training data. Our work
compromises by only detecting the most common events,
which already reduces the mental burden on streamers,
as they no longer have to comment on every single kill or
destroyed turret. Future work could expand upon this by
improving data collection eforts to include a wider range
of in-game events. As our work was only preliminary
in nature, we have not performed a thorough evaluation
with human viewers at this point.</p>
      <p>With results on the human perception of the generated
commentary, it will also be possible to do more research
to determine the optimal frequency of commentary
delivery to maintain entertainment value while conveying
significant insights, as well as how the system can best
interact with the streamer’s own commentary.
Furthermore, while the paper focuses on visual events, many
games have non-visual events such as sound cues for
horror games, which could also be included in the event
detection.</p>
    </sec>
    <sec id="sec-5">
      <title>7. Acknowledgements</title>
      <p>We would like to thank our reviewers, in particular for
their suggestions to improve our section on ethical
considerations. We believe this discussion is essential for
work such as this, and the feedback that was provided
has been invaluable.
r_a_slain
b_e_slain
r_a_slain
b_e_slain
r_a_slain</p>
      <p>Generated Commentary
"That’s it! I’ve had enough! I’m gonna show
them who’s boss! Time to take down this
enemy team for good!"
"What the heck?! That enemy team just got
a double kill! I can’t believe it! They must be
pretty good! I better watch out for them!"
Wubba lubba dub dub! We got a kill! That’s
gonna be a big score for our team! Let’s keep
it up!
"Oh great, the enemy team got a kill! This
is gonna be a tough match for us! We need
to stay focused and work together if we’re
gonna have a chance of winning!"
That’s it, I’m done! Those no-good,
lowdown, dirty scoundrels just got another kill!
I’m outta here!
"Hah! Looks like the enemy team got SHUT
DOWN! That’s what happens when you
mess with Rick Sanchez!"
"Oh great, the enemy team just got a kill!
That’s just what I need, more trouble for me
to deal with. I’m gonna have to step up my
game if I want to get my team back in the
game.
"Oh, come on! That was a cheap shot! You
didn’t have to go and get that kill!"
"Oh great, they got a kill! Looks like I’m
gonna have to step up my game if I’m gonna
take them down!"
"Geez Louise! That was a close one. Looks
like the enemy team got the kill this time.</p>
      <p>I guess we’ll just have to try harder next
time!"
b_e_slain
r_double
defeat
"Oh man, that enemy team is really on fire!
They just got a double kill and it looks like
they’re not slowing down anytime soon!
This is gonna be a tough fight!"
WUBBA LUBBA DUB DUB! That was a
great play by my team! We got that kill
and now we’re one step closer to victory!
"What?! That’s incredible! They must have
worked together as a team to pull of that
play. I’m impressed!"
"Oh man, that enemy team is really playing
well! They just got a double kill and are
looking pretty unstoppable right now. We’re
gonna have to step up our game if we want
to take the win!"
Oh man, that’s a tough one! Those guys
must be really good players. I guess I’ll have
to step up my game if I want to win this
match!
Wubba Lubba Dub Dub! That was an
awesome kill by my team! Can’t wait to see
what our next move will be!
Oh man, my team just got a kill! The enemy
team didn’t stand a chance! Wubba lubba
dub dub!
That was a real shame! Looks like our
opponents got the better of us this time. We
need to step up our game and make sure we
don’t let this happen again.</p>
      <p>Wubba Lubba Dub Dub! That was a great
play by my team! We’re step closer to
victory!
"Oh great, just great! It looks like the enemy
team had the upper hand this round. That’s
just what I need, another loss on my record.</p>
      <p>I’m gonna have to put in some extra practice
if I want to come</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Obrist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wright</surname>
          </string-name>
          ,
          <article-title>Live-streaming changes the (video) game</article-title>
          , in
          <source>: Proceedings of the 11th european conference on Interactive TV and video</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>131</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Ferguson</surname>
          </string-name>
          ,
          <article-title>Sports announcer talk: Syntactic aspects of register variation</article-title>
          ,
          <source>Language in society 12</source>
          (
          <year>1983</year>
          )
          <fpage>153</fpage>
          -
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Guzdial</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Riedl</surname>
          </string-name>
          ,
          <source>Automated let's play commentary</source>
          , arXiv preprint arXiv:
          <year>1909</year>
          .
          <volume>02195</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gandhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Harrison</surname>
          </string-name>
          ,
          <article-title>End-to-end let's play commentary generation using multi-modal video representations</article-title>
          ,
          <source>in: Proceedings of the 14th International Conference on the Foundations of Digital Games</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Voelz</surname>
          </string-name>
          , E. André, G. Herzog, T. Rist,
          <article-title>Rocco: A robocup soccer commentator system</article-title>
          ,
          <source>in: RoboCup98: Robot Soccer World Cup II 2</source>
          , Springer,
          <year>1999</year>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>60</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Tanaka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nakashima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Noda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hasida</surname>
          </string-name>
          , I. Frank,
          <string-name>
            <given-names>H.</given-names>
            <surname>Matsubara</surname>
          </string-name>
          ,
          <string-name>
            <surname>Mike:</surname>
          </string-name>
          <article-title>An automatic commentary system for soccer</article-title>
          ,
          <source>in: Proceedings International Conference on Multi Agent Systems (Cat. No. 98EX160)</source>
          , IEEE,
          <year>1998</year>
          , pp.
          <fpage>285</fpage>
          -
          <lpage>292</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Binsted</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Luke</surname>
          </string-name>
          ,
          <article-title>Character design for soccer commentary</article-title>
          , in: RoboCup-98
          <source>: Robot Soccer World Cup II 2</source>
          , Springer,
          <year>1999</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <article-title>Automatic baseball commentary generation using deep learning</article-title>
          ,
          <source>in: Proceedings of the 35th Annual ACM Symposium on Applied Computing</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1056</fpage>
          -
          <lpage>1065</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Riot</given-names>
            <surname>Games</surname>
          </string-name>
          , League of legends [PC game], Santa Monica, CA,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Floridi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chiriatti</surname>
          </string-name>
          , Gpt-3
          <article-title>: Its nature, scope, limits, and consequences</article-title>
          ,
          <source>Minds and Machines</source>
          <volume>30</volume>
          (
          <year>2020</year>
          )
          <fpage>681</fpage>
          -
          <lpage>694</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <article-title>Table 3 Generated commentary for table 1 L.</article-title>
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          <string-name>
            <surname>Gomez</surname>
            , Ł. Kaiser,
            <given-names>I. Polosukhin</given-names>
          </string-name>
          , At- ability, and transparency,
          <year>2021</year>
          , pp.
          <fpage>610</fpage>
          -
          <lpage>623</lpage>
          .
          <article-title>tention is all you need</article-title>
          ,
          <source>Advances in neural infor-</source>
          [16]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Bai</surname>
          </string-name>
          , M. Lyu,
          <source>mation processing systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
          <article-title>Biasasker: Measuring the bias in conversational ai</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Taveekitworachai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Abdullah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Dewantoro</surname>
          </string-name>
          , system,
          <source>arXiv preprint arXiv:2305.12434</source>
          (
          <year>2023</year>
          ). R. Thawonmas,
          <string-name>
            <given-names>J.</given-names>
            <surname>Togelius</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Renz,</surname>
          </string-name>
          <article-title>Chatgpt4pcg competition: Character-like level generation for science birds</article-title>
          ,
          <source>arXiv preprint arXiv:2303.15662</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <article-title>Eficientdet: Scalable and eficient object detection</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>10781</fpage>
          -
          <lpage>10790</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B.</given-names>
            <surname>Koonce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Koonce</surname>
          </string-name>
          , Eficientnet,
          <article-title>Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization (</article-title>
          <year>2021</year>
          )
          <fpage>109</fpage>
          -
          <lpage>123</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Bender</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gebru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>McMillan-Major</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shmitchell</surname>
          </string-name>
          ,
          <article-title>On the dangers of stochastic parrots: Can language models be too big?</article-title>
          ,
          <source>in: Proceedings of the 2021 ACM conference on fairness, account-</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>