<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Bringing Rome to Life: Evaluating Historical Image Generation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Phillip B.Ströbel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zejie Guo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ülkü Karagöz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eva MariaWill</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Felix K.Maier</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CHR 2024: Computational Humanities Research Conference</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computational Linguistics, University of Zurich</institution>
          ,
          <addr-line>Andreasstrasse 15, 8050 Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of History, University of Zurich</institution>
          ,
          <addr-line>Karl Schmid-Strasse 4, 8006 Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <fpage>113</fpage>
      <lpage>126</lpage>
      <abstract>
        <p>This study evaluates the potential of AI image generation for visualising historical events, focusing on two ancient Roman scenarios: the Roman triumph and theLupercalia festival. Using DALL-E 3, we generated 600 images based on 100 prompts derived from scientific texts. We then conducted a twopart evaluation: (1) A human evaluation by 21 history students, who compared image pairs and rated individual images on accuracy and prompt alignment, and (2) two automated analyses, one modelled after the human evaluation protocol and one using visual question-answering (VQA) techniques. Our results reveal both the promise and limitations of AI in historical visualisation. While DALL-E 3 produced many convincing images, there were notable discrepancies between human and automated assessments. We found that Large Language Models tend to rate images more favourably than human evaluators. We contribute a novel dataset for historical image generation, initial human and automated evaluation protocols, and insights into the challenges of using AI for historical visualisation, which is incredibly important for historians to reconstruct past events. Our findings highlight the need for refined evaluation methods and underscore the complexity of assessing historical accuracy in AI-generated imagery. This study lays the groundwork for future research on improving AI models for historical visualisation and developing more robust evaluation frameworks.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Digital Humanities</kwd>
        <kwd>image generation</kwd>
        <kwd>human evaluation</kwd>
        <kwd>automatic evaluation</kwd>
        <kwd>history</kwd>
        <kwd>image dataset</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Historians, akin to criminologists, analyse primary sources and eyewitness accounts to extract
meaning and understand the motives and circumstances of historical events. However, unlike
criminologists, who can re-enact events, historians face the challenge of studying occurrences
that cannot be replicated or reproduced in experiments. This presents a significant challenge
in their work.</p>
      <p>Criminologists have developed methods to mitigate the uncertainties involved. Re-enacting
crucial moments of an action or crime using real people or AI-based simulations has become
one of the most potent tools in criminology. These re-enactments allow us to visualise the
action, providing a cinematic perspective that clarifies a crime’s sequence and spatial dynamics.
This process enhances our understanding by enabling us to perceive previously unseen details
and prompting further questions. By experiencing events as if we were witnesses, we gain a
newfound clarity.</p>
      <p>Surprisingly, despite the similar challenges faced by both professions, historians have yet to
embrace this re-enactment approach fully. Our project aims to change that. Leveraging rapid
advancements in AI development, we aim to introduce an innovative platform to redefine how
we perceive and comprehend historical events. Our goal is to develop a web application that
generates storyboards or individual images from historical texts.</p>
      <p>This ‘re-experiencing’ of history will empower users to recapture seemingly ‘lost’
historical moments. Users can model specific actions or occasions from diverse perspectives by
employing various scenarios with AI support. This approach will uncover performative
dynamics, potentially revealing previously undisclosed aspects of historical events, much like
the re-enactment in criminology.</p>
      <p>However, we must question the suitability of such image-generation models. They require
testing for historical accuracy before we can employ them for the previously mentioned
purposes. We chose two specific Roman scenarios to test the capacity of DALL-E 3 to create
historically accurate and engaging images: the Roman procession and tLhuepercalia festival.</p>
      <sec id="sec-1-1">
        <title>1.1. Our Contribution</title>
        <p>Our study focuses on these two events due to their significance in Roman culture and the
varying levels of textual and visual documentation available for each. The Roman triumph, a
well-documented celebration of military victory, provides a rich base of textual descriptions.
In contrast, the Lupercalia, an ancient fertility festival, ofers a more challenging scenario with
fewer detailed contemporary accounts.</p>
        <p>To assess DALL-E 3’s capabilities in this domain, we generated 600 images – 450 for the
triumph and 150 for theLupercalia (see Section 3). Our evaluation process is twofold:
1. Human evaluation: We conducted a comprehensive review involving 21 advanced
history students to assess the images’ historical accuracy.
2. Automated analysis: We employed computer vision techniques to analyse the images
for prompt alignment.</p>
        <p>This dual approach allows us to measure the generated images’ subjective impact on human
viewers and their objective alignment with historical data. Our research contributes to the
broader discussion of AI’s potential in historical visualisation and its limitations and contains
the following items:
1. A novel, automatically generated dataset comprising 100 prompts and 600 images for
historical image generation.
2. An initial human evaluation of a subset of these automatically generated images.
3. An initial automatic evaluation of the same subset.
4. An assessment of how well human and automatic evaluation correlate.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        The evaluation of automatically generated images has recently gained traction, mainly due
to the increasingly sophisticated image generation models. Otani, Togashi, Sawai, Ishigami,
Nakashima, Rahtu, Heikkilä, and Satoh 2[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] contemplated, based on an extensive analysis of
37 papers, that human evaluation protocols are often not reproducible and lack a clear
description. Moreover, evaluation usually relies on automatic measures that poorly align with human
scores.
      </p>
      <p>
        The advantage of human feedback is that it can improve text-to-image models, e.g., with
reinforcement learning from human feedback (as used in Natural Language Processi2n8g])[. Xu,
Liu, Wu, Tong, Li, Ding, Tang, and Dong 3[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] exploited a dataset of 8,878 prompts and 136,892
image comparisons to fine-tune a reward model that aligns more closely with human
preferences. Liang, He, Li, Li, Klimovskiy, Carolan, Sun, Pont-Tuset, Young, Yang, Ke, Dvijotham,
Collins, Luo, Li, Kohlhof, Ramachandran, and Navalpakkam1[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] used human feedback
concerning Plausibility, Aesthetics, Text-image Alignment, and an Overall impression to predict
human feedback scores. Due to the successful integration of human feedback in the model
ifne-tuning by Xu, Liu, Wu, Tong, Li, Ding, Tang, and Dong [
        <xref ref-type="bibr" rid="ref32">33</xref>
        ], we created an evaluation
scenario which allows us to integrate such feedback directly in future work (see Sec4ti.o1)n.
      </p>
      <p>
        While Xu, Liu, Wu, Tong, Li, Ding, Tang, and Dong3[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] focused on prompt-to-image
alignment, other image properties are open for evaluation. Lee, Yasunaga, Meng, Mai, Park, Gupta,
Zhang, Narayanan, Teufel, Bellagente, Kang, Park, Leskovec, Zhu, Li, Wu, Ermon, and Liang
[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] worked on holistic image evaluation and identified twelve aspects among which we find
Alignment, Quality, Aesthetics, and Originality (among others). Evaluating each aspect calls for
diferent measures, some of them human, some of them automated. They created a holistic
image evaluation benchmark for existing datasets and reported scores for all aspects and 26
models. While such an evaluation efort is valuable and provides a helpful oversight, we focus
on prompt-to-image alignment evaluation in this work.
      </p>
      <p>The research mentioned above has had access to large and heterogeneous datasets and
results from extensive evaluation campaigns. In the context of historical image generation, such
work does not yet exist. One exception is the investigation of Fareed, Bou Nassif, and Nofa8l] [
who tested the usage of Leonardo1 for teaching purposes in the field of “History of
Architecture”. They evaluated the usability of Leonardo with a questionnaire after a workshop, which
generally showed a need for the evaluation of AI-generated images for usage in the historical
domain.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Data Collection with DALL-E 3</title>
      <p>
        Next, we outline the methodology for data collection using DALL-E 3 to generate images
related to triumphal processions and thLeupercalia, which included the following steps:
1. Collecting Historical Documents: We collected resources (i.e., academic papers,
books, and other relevant documents) about the triumph and thLeupercalia in ancient
1See https://leonardo.a.i
Rome. Specifically, we included five documents related to theLupercalia [
        <xref ref-type="bibr" rid="ref10 ref20 ref28 ref31 ref7">32, 29, 20, 7,
10</xref>
        ] and 15 documents focused on triumphal procession2s7[
        <xref ref-type="bibr" rid="ref1 ref12 ref13 ref14 ref15 ref18 ref19 ref2 ref22 ref24 ref29 ref3 ref9">, 22, 3, 2, 15, 14, 18, 23, 12,
25, 13, 19, 9, 1, 30</xref>
        ].
2. Creating Prompts from Documents: For each document, we manually derived five
prompts. Each prompt was designed to capture a specific scene described in the texts.
E.g., a document on triumphal processions could include prompts about the attire
Romans wore, the types of vehicles used, or the procession sequence. In total, we created
100 prompts.
3. Image Generation with DALL-E 3: We used each prompt to generate six images using
DALL-E 3 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] via the OpenAI API.2 The 100 prompts resulted in 150 generated images
for theLupercalia and 450 for the triumphal procession3s.
      </p>
      <p>Note that we did not force the model to produce realistic images. This led to a great variety
of image styles, some of which are indeed life-like, while others are more in the style of a
Renaissance painting or a black-and-white pencil sketch. All prompts, however, are based on
scientific literature. See Figure1 for example images and prompts from the datase4t.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluating Automatically Generated Data</title>
      <p>The following sections focus on the diferent evaluation scenarios employing human
annotators and automatic evaluation measures.</p>
      <sec id="sec-4-1">
        <title>4.1. Human Evaluation</title>
        <sec id="sec-4-1-1">
          <title>4.1.1. Human Evaluation Setup</title>
          <p>
            We generated two evaluation scenarios to obtain feedback from human annotators.
Image Comparison (IC) The first scenario asks annotators to decide which of two images
better reflects the prompt. This is a cognitively easier task. Much in the manner of Xu, Liu,
Wu, Tong, Li, Ding, Tang, and Dong [
            <xref ref-type="bibr" rid="ref32">33</xref>
            ], we plan to use these ratings for fine-tuning models to
produce more faithful images. The participants are instructed not to judge the image style. We
(−1)
only compared images generated with the same prompt, which, based on the formula2
to find unique pairings, results in 15 pairs per prompt (as mentioned in the previous section,
we generated six images per prompt). Multiplied with the 100 total prompts in the dataset, we
arrive at 1,500 comparisons.
          </p>
          <p>Image Rating (IR) The second task requires the participants to rate an image on a 5-point
Likert scale with the following options:</p>
          <p>1. The image does not match the prompt at all.
2See https://openai.com/index/openai-ap.i
3The image generation costs amount to $48.06.
4The whole dataset (images and prompts) is available on GitHub. Sehettps://github.com/AncientHistory-UZH/C
HR2024_prompt-and-image-datase.t</p>
          <p>2. The image barely contains aspects of the prompt.
3. The image catches some aspects of the prompt, but it is not very accurate.
4. The image catches most of the aspects of the prompt.</p>
          <p>5. The image completely matches the prompt.</p>
          <p>Additionally, we asked the users to describe which aspects of the image dniodt correspond to
the prompt in a text field. In this scenario, which demands more time and efort, we need 600
ratings for one complete dataset annotation.</p>
          <p>We set up a Prodigy interface5, which we used to obtain the assessment of the annotators.
See Figure2 to get an impression of the annotation environment. We recruited 21 advanced
history students for the annotations. We did not ask the participants to annotate a specific
number of pairs. They were compensated with book vouchers of a value of $30. An online
meeting was organised to explain the guidelines, emphasising that in the first scenario, they
should judge based on the alignment of the images with the prompts rather than their visual
appeal. They should consider visual features only if the two images reflect the prompts equally.
The students spent approximately one afternoon annotating the data in both scenarios.
5See https://prodi.gy.</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>4.1.2. Results of Human Evaluation</title>
          <p>Table 1 gives an overview of the results from the human evaluation. In the IC setting, we
received 1,569 comparisons. 103 samples were annotated more than once. For unknown
reasons, 64 data points did not contain the human assessment, so we excluded them from further
analysis. On average, each participant compared 74.71 (SD 43.32) image pairs.</p>
          <p>The IR scenario received less feedback since the participants provided written feedback in
a text field besides their rating. We obtained 568 ratings, of which 29 were double ratings—24
feedbacks without scores needed to be excluded.</p>
          <p>We must note here that, due to a wrong parameter setting of Prodigy in both scenarios, the
data samples to be evaluated were presented to the participants in sequential instead of a
random order. This led to only marginal annotation overlap. For this reason, we cannot compute
inter-annotator agreements (IAA) yet. However, since we plan to improve the models with the
feedback obtained from the participants, we will have further evaluation rounds during which
we can take care of this limitation. Still, to the best of our knowledge, this is the first
“largescale” evaluation campaign dedicated to historical image generation. We can still analyse and
compare the results obtained with the limitations in mind (see Sectio4.n1.3).</p>
          <p>However, since previous studies reported low IAA in human evaluation scenarios 1(c6f].)[,
we hypothesise a similar outcome on our dataset.</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>4.1.3. Comparison of Human Results with Large Language Model (LLM) Evaluation</title>
          <p>
            To mitigate the missing information on IAA and to evaluate the suitability of multimodal LLMs
for scoring tasks, we employed GPT-4o2[
            <xref ref-type="bibr" rid="ref1">1</xref>
            ], Gemini 1.5 Pro [
            <xref ref-type="bibr" rid="ref25">26</xref>
            ] and Claude 3.5 Sonnet6. We let
the LLMs solve the same tasks as the annotators, i.e., we applied them to the IC (only GPT-4o)
and IR (all three) evaluation scenari o7s.
          </p>
          <p>For IC, Table2 shows the agreements of the human comparisons with GPT-4o’s comparisons.
We see that in 57.51% of the cases, human annotators and GPT-4o agree on which of the two
images better corresponds to the prompt.</p>
          <p>Figure 3 summarises the results for the IR setting. The left graph shows the diferences
between the human and the LLM ratings. The tendency is that LLMs rate images higher than
human annotators. The right graph shows the LLM’s deviations from the human scores. E.g.,
in 164 (30.15%) ratings, GPT-4o agrees with the human scores. In 169 (31.07%) cases,
GPT4o scores one point higher on the Likert scale than the human annotators (i.e., GPT-4o had
rated an image a 3 when the human annotator rated it at 2). We see that Claude tends to rate
images higher, especially. Overall, the deviations seem normally distributed, a fact that might
be exploited for future evaluations.</p>
          <p>Choosing two scenarios to evaluate allows us to test for diferences in assessing images
6See https://www.anthropic.com/news/claude-3-5-sonne.t
7This generated costs of $19.14 for GPT-4o, $2.49 for Gemini and $5.27 for Claude.
Inter-annotator agreement between diferent groups using Krippendorf’s alpha. We used the same 544
images for which we have computed the t-test..</p>
          <p>GPT vs. Gemini vs. Claude</p>
          <p>GPT vs. human</p>
          <p>Triumph
between the triumph and theLupercalia scenario. Our null hypothes is0 is that there is no
diference in the ratings of human annotators and, e.g., GPT-4o in the two historical scenarios.
variation and (ii) unequal sample sizes. For the human evaluation (unifying the assessment
results but excluding invalid samples), thpe-value does not allow us to reject0. The GPT
and Gemini ratings show another picture. Thpe-values show a highly significant diference
between ratings of the triumph and theLupercalia images. Claude’sp-value is on the brink of
showing a statistically significant diference. The, on average, lower ratings by LLMs of the
Lupercalia images could indicate DALL-E’s difÏculties in generating adequate imagery. Firstly,
since the Lupercalia are not so much a described nor illustrated phenomenon, it is reasonable
that images portraying the festival are not on the same standard as those generated for the
triumphal procession. Secondly, the automatic evaluation poses problems for LLMs because
they do not “know” as much as they do for the triumph.</p>
          <p>Although we cannot provide IAA scores for the human evaluation yet, we can do so for the
automatically generated ratings by the LLMs. Tabl4e shows the results when we compare
the ratings for the LLMs (again split into triumph- anLdupercalia-related scores). The scores
are all around 0, indicating low overlap, IAA. Unifying all human scores and comparing them
against the ratings obtained via GPT-4o also shows low overlap. These results hint at the very
diferent rating “strategies” of the LLMs. We need further evaluation to shed more light on the
origins of the discrepancies.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Automatic Evaluation</title>
        <sec id="sec-4-2-1">
          <title>4.2.1. Automatic Evaluation Setup</title>
          <p>
            For a further fully automatic evaluation procedure, we employed the Question Generation and
Answering (QG/A) [
            <xref ref-type="bibr" rid="ref11 ref6">11, 6</xref>
            ] framework for automatic image evaluation. The first step in this
framework involves using a pre-trained language model to generate a set of questions based
on a given prompt and question-generation instructions via few-shot learning. In the second
step, a pre-trained multimodal model generates answers given the image and the generated set
of questions.
          </p>
          <p>
            Question Generation (QG) In our study, we utilised GPT-3.55[] for QG employing the
Davidsonian Scene Graph (DSG) [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ] method. DSG serves as an evaluation framework grounded
in formal semantics. This method’s main advantage is its ability to generate atomic and unique
questions structured in dependency graphs, which (i) ensure comprehensive semantic
coverage and (ii) avoid inconsistencies in responses. Cho, Hu, Garg, Anderson, Krishna, Baldridge,
Bansal, Pont-Tuset, and Wang [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ] empirically demonstrated that DSG addresses the challenges
of hallucinations, duplications, and omissions in QG.
          </p>
          <p>Visual Question Answering (VQA) We employed GPT-4o for the VQA task. The following
prompt instruction guides the model: “You are a helpful assistant. Please answer the question
only with ‘Yes’ or ‘No’. Do not give other outputs. Questio{nq:uestion}.” To ensure precise
control over the output, specifically responding with either ‘Yes’ or ‘No’, we set the parameter
logit_bias to 100 for both ‘Yes’ and ‘No’ tokens. Logit bias modifies the likelihood of
speciifed tokens appearing in the model-generated output. We also set thetop_p (nucleus sampling)
parameter to 0.1 to restrict the model’s consideration to a subset of tokens (the nucleus) whose
cumulative probability mass reaches a designated threshold (top-p). In the context of a 0.1
top_p setting, the model exclusively considers tokens constituting the top 10% of the
probability mass for the subsequent token. The combination olfogit_bias and top_p configurations
enables the outputs to adhere to predefined patterns (‘Yes’ and ‘No’), rendering the model more
deterministic and particularly suitable for our image evaluation8taWske. assign a score of 1
for ‘Yes’ and 0 for ‘No’ and then compute an average score for each image. We observe that
GPT occasionally generates questions such as “Is there an image?” or “Can you visualize a
scene?” which are invalid in our context, as the input consistently includes an image and a set
of questions. We excluded the scores of these invalid questions from our analysis.</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Results of VQA</title>
          <p>Figure4 shows a histogram of the results of the VQA scores for all 600 images. We find most
scores between 0.5 and 0.9, with over 60 images obtaining a perfect score of 1. This means that
each ‘Yes-or-No’ question was answered with ‘Yes’. When we look at three results as presented
in Figure5 in Appendix A, we find that VQA attributes a low score of 0.05 for imagea). The
human evaluator and GPT, however, have scored this image with a 4 in the IR scenario. bI)n,
we have a medium VQA score of 0.61, a human score of 5 and a GPT score of 4. Lastlcy),shows
an image with a VQA score of 1, but a human annotator scored this image a 3 and GPT a 4.
We already see discrepancies between the diferent scores from these three examples only. A
comparison of VQA between the 450 images from the triumphal procession and the 150 images
8This evaluation scenario cost us $7.69. The whole experiment, i.e., image generation, LLM evaluation in the two
scenarios from Section4.1.3 and the one mentioned in this section totalled at $80.67.
from theLupercalia based on Welch’s t-test shows no significant diferences between the two
ratings ( = 0.88 ). From this, we conclude that ratings based on VQA produce more reliable
results than those produced with a Likert scale.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Limitations and Outlook</title>
      <p>The most significant limitation of our work is the missing IAA scores. For future evaluation
rounds, we will set up the evaluation to allow for their computation. In this way, we get
reliable measures of how demanding the task of assessing the alignment of historical images
and the prompts they produced is. However, we argue that the results we obtained from the
human evaluation are still valuable and allow for fine-tuning models based on human feedback
(preferences in the IC and textual input in the IR scenario), albeit in a low-resource setting.</p>
      <p>Moreover, we will employ more models to generate images in future experiments. This
approach enables us to decide which models are the most suitable for historical image
generation. The stable prompt base also allows for comparable results. Still, the significant number
of images we will generate in future endeavours also calls for automatic evaluation methods.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In conclusion, our study provides valuable insights into the potential and challenges of using
AI for historical image generation. The evaluation of 600 AI-generated images of triumphal
processions and theLupercalia revealed both promising capabilities and significant limitations.</p>
      <p>Our findings hint at the discrepancies between human and automated assessments,
underscoring the complexity of evaluating historical accuracy in AI-generated imagery. Ultimately,
this study serves as a stepping stone towards more sophisticated use of AI in historical
recreation and education while cautioning against over-reliance on automated systems for historical
interpretation.</p>
      <p>This research contributes a novel dataset and evaluation framework to the field, enabling
future studies. As AI continues to evolve, our work suggests that while it holds promise for
enhancing historical visualisation and understanding, it requires careful human oversight and
interpretation.
[23] I. Östenberg. “Triumph and spectacle. Victory celebrations in the Late Republican civil
wars”. In: The Roman Republican Triumph Beyond the Spectacle. 2014, pp. 181–193.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Algül</surname>
          </string-name>
          . “
          <article-title>The Roman Triumph: Participation, Historiography and Remembrance”</article-title>
          . In: (
          <year>2018</year>
          ). url: https://www.academia.
          <source>edu/43295099/The%5C%5FRoman%5C%5FTriumph %5C%5FParticipation%5C%5FHistoriography%5C%5Fand%5C%5FRemembran.ce</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Armstrong</surname>
          </string-name>
          . “Claiming Victory:
          <article-title>The Early Roman Triumph”</article-title>
          .
          <source>ISnp:alinger, Anthony John</source>
          <year>1947</year>
          -
          <article-title>(edt), Armstrong, Jeremy (edt), Rituals of triumph in the Mediterranean world</article-title>
          .
          <source>Culture and History of the Ancient Near East</source>
          <volume>63</volume>
          . Leiden u.a.:
          <source>Brill</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>7</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Beard</surname>
          </string-name>
          .
          <article-title>The Roman Triumph</article-title>
          . Harvard University Press,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Betker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Goh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Brooks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ouyang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Guo</surname>
          </string-name>
          , et al. “
          <article-title>Improving Image Generation with Better Captions”</article-title>
          . In: (
          <year>2023</year>
          ). urlh:ttps://cdn.o penai.com/papers/dall-e-3.pd. f
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herbert-Voss</surname>
          </string-name>
          , G. Krueger,
          <string-name>
            <given-names>T.</given-names>
            <surname>Henighan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Child</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ziegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Winter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hesse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          , E. Sigler,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chess</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Berner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>McCandlish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          . “
          <article-title>Language Models are Few-Shot Learners”</article-title>
          .
          <source>InA:dvances in Neural Information Processing Systems</source>
          . Ed. by
          <string-name>
            <given-names>H.</given-names>
            <surname>Larochelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ranzato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Balcan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          . Vol.
          <volume>33</volume>
          . Curran Associates, Inc.,
          <year>2020</year>
          , pp.
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          . urlh: ttps://proceedings.neurips.c c/paper%5C%5Ffiles/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.p.df
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Garg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Krishna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baldridge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pont-Tuset</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          . “
          <article-title>Davidsonian Scene Graph: Improving Reliability in Fine-grained Evaluation for Text-to-Image Generation”</article-title>
          . In: (
          <year>2023</year>
          ). urlh:ttp://arxiv.org/abs/2310.18235.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Š</surname>
          </string-name>
          . Erker. “
          <article-title>Das Lupercalia-Fest im augusteischen Rom: Performativität, Raum und Zeit”</article-title>
          .
          <source>In: Archiv für Religionsgeschichte 11.1</source>
          (
          <issue>2009</issue>
          ), pp.
          <fpage>145</fpage>
          -
          <lpage>178</lpage>
          . doi: doi:10.1515/97831102089 62.2.145.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Fareed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Bou</given-names>
            <surname>Nassif</surname>
          </string-name>
          , and
          <string-name>
            <surname>E. Nofal.</surname>
          </string-name>
          “
          <article-title>Exploring the Potentials of Artificial Intelligence Image Generators for Educating the History of Architecture”</article-title>
          .
          <source>HIne:ritage 7</source>
          .3 (
          <issue>2024</issue>
          ), pp.
          <fpage>1727</fpage>
          -
          <lpage>1753</lpage>
          . doi:
          <volume>10</volume>
          .3390/heritage7030081.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9] “
          <article-title>Der römische Triumph in Prinzipat und Spätantike: Probleme - Paradigmen - Perspektiven”</article-title>
          . In: Der römische Triumph in Prinzipat und Spätantike. Ed. by
          <string-name>
            <given-names>F.</given-names>
            <surname>Goldbeck</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Wienand</surname>
          </string-name>
          . Berlin, Boston: De Gruyter,
          <year>2017</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          . dodi:oi:
          <volume>10</volume>
          .1515/9783110448009-0
          <fpage>03</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Guarisco</surname>
          </string-name>
          . “
          <article-title>Augustus, the Lupercalia and the Roman identity”</article-title>
          .
          <source>InA:cta Antiqua Academiae Scientiarum Hungaricae</source>
          <volume>55</volume>
          .
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          (
          <year>2015</year>
          ), pp.
          <fpage>223</fpage>
          -
          <lpage>228</lpage>
          . doi:
          <volume>10</volume>
          . 1556 / 068 .
          <year>2015</year>
          .
          <volume>55</volume>
          .1-
          <fpage>4</fpage>
          .16. url: https://akjournals.com/view/journals/068/55/1-4/article-p223.
          <article-title>x</article-title>
          .ml
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kasai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Krishna</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          . “TIFA:
          <article-title>Accurate and Interpretable Text-to-Image Faithfulness Evaluation with Question Answering”</article-title>
          . In: (
          <year>2023</year>
          ). url: http://arxiv.org/abs/2303.11897.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lange</surname>
          </string-name>
          . “
          <article-title>Mock the Triumph: Cassius Dio, Triumph and Triumph-Like Celebrations”</article-title>
          .
          <source>In: Cassius Dio. Brill's Historiography of Rome and Its Empire Series. Brill</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>92</fpage>
          -
          <lpage>114</lpage>
          . doi:
          <volume>10</volume>
          .1163/9789004335318\_
          <fpage>007</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Lange</surname>
          </string-name>
          . “
          <article-title>The Late Republican Triumph: Continuity and Change”</article-title>
          .
          <source>InD:er römische Triumph in Prinzipat und Spätantike</source>
          . Berlin, Boston: De Gruyter,
          <year>2017</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>58</lpage>
          . doi: doi:10.1515/
          <fpage>9783110448009</fpage>
          -
          <lpage>004</lpage>
          . url: https://doi.org/10.1515/
          <fpage>9783110448009</fpage>
          -
          <lpage>00</lpage>
          4.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lange</surname>
          </string-name>
          . “
          <article-title>The Triumph outside the City: Voices of Protest in the Middle Republic”</article-title>
          . In: The Roman Republican Triumph. Ed. by
          <string-name>
            <given-names>C.</given-names>
            <surname>Hjort Lange</surname>
          </string-name>
          and
          <string-name>
            <given-names>F.</given-names>
            <surname>Vervaet</surname>
          </string-name>
          . Analecta Romana Instituti,
          <source>Suppl. Quasar</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>67</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Lange</surname>
          </string-name>
          . “
          <article-title>Triumph and Civil War in the Late Republic”</article-title>
          .
          <source>InPa:pers of the British School at Rome</source>
          <volume>81</volume>
          (
          <year>2013</year>
          ), pp.
          <fpage>67</fpage>
          -
          <lpage>90</lpage>
          . doi:
          <volume>10</volume>
          .1017/s0068246213000056.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yasunaga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Narayanan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Teufel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bellagente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.-F.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ermon</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Liang</surname>
          </string-name>
          . “
          <article-title>Holistic Evaluation of Text-to-Image Models”</article-title>
          .
          <source>IAn:dvances in Neural Information Processing Systems</source>
          . Ed. by
          <string-name>
            <given-names>A.</given-names>
            <surname>Oh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Globerson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Saenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Levine</surname>
          </string-name>
          . Vol.
          <volume>36</volume>
          . Curran Associates, Inc.,
          <year>2023</year>
          , pp.
          <fpage>69981</fpage>
          -
          <lpage>70011</lpage>
          . urhl:ttps: //proceedings.neurips.cc/paper%5C%5Ffiles/paper/2023/file/dd83eada2c3c74db3c7fe1c0 87513756-Paper-Datasets%
          <article-title>5C%5Fand%5C%5FBenchmarks.pd</article-title>
          .f
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Klimovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Carolan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pont-Tuset</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Dvijotham</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Collins</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Luo</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>K. J.</given-names>
          </string-name>
          <string-name>
            <surname>Kohlhof</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramachandran</surname>
            , and
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Navalpakkam</surname>
          </string-name>
          . “
          <article-title>Rich Human Feedback for Text-to-Image Generation”</article-title>
          .
          <source>InP:roceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          .
          <year>2024</year>
          , pp.
          <fpage>19401</fpage>
          -
          <lpage>19411</lpage>
          . url: https://openaccess.thecvf.com/content/CVPR2024/papers/Liang %5C%
          <article-title>5FRich%5C%5FHuman%5C%5FFeedback%5C%5Ffor%5C%5FText-to-Image%5C%5 FGeneration%5C%5FCVPR%5C%5F2024%5C%5Fpaper.pd</article-title>
          .f
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Madsen</surname>
          </string-name>
          . “
          <article-title>The Loser's Prize: Roman Triumphs and Political Strategies during the Mithridatic Wars”</article-title>
          .
          <source>In:The Roman Republican Triumph Beyond the Spectacle. Analecta Romana Instituti Danici. Supplementa Xlv. Quasar</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>117</fpage>
          -
          <lpage>130</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Mittag</surname>
          </string-name>
          . “
          <article-title>Die Triumphatordarstellung auf Münzen und Medaillons in Prinzipat und Spätantike”</article-title>
          .
          <source>In: Der römische Triumph in Prinzipat und Spätantike</source>
          . Berlin, Boston: De Gruyter,
          <year>2017</year>
          , pp.
          <fpage>419</fpage>
          -
          <lpage>452</lpage>
          . doi:doi:10.1515/
          <fpage>9783110448009</fpage>
          -
          <lpage>017</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>J. A. North. “</surname>
          </string-name>
          <article-title>Caesar at the Lupercalia”</article-title>
          .
          <source>InJ:ournal of Roman Studies</source>
          <volume>98</volume>
          (
          <year>2008</year>
          ), pp.
          <fpage>144</fpage>
          -
          <lpage>160</lpage>
          . doi:
          <volume>10</volume>
          .3815/007543508786239210.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>OpenAI. Hello</surname>
          </string-name>
          GPT-4o.
          <year>2024</year>
          . url: https://openai.com/index/hello-gpt-4.o
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>I.</given-names>
            <surname>Östenberg</surname>
          </string-name>
          . Staging the World: Spoils, Captives, and
          <article-title>Representations in the Roman Triumphal Procession</article-title>
          . Oxford: Oxford University Press,
          <year>2009</year>
          . do1i:
          <fpage>0</fpage>
          .1093/acprof:oso/9780 199215973.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Otani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Togashi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sawai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ishigami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Nakashima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Rahtu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Heikkilä</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Satoh</surname>
          </string-name>
          . “
          <article-title>Toward Verifiable and Reproducible Human Evaluation for Text-to-Image Generation”</article-title>
          .
          <source>In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          .
          <year>2023</year>
          , pp.
          <fpage>14277</fpage>
          -
          <lpage>14286</lpage>
          . url: https://cvpr2023.thecvf.com/virtual/2023/poster/220 14.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Popkin</surname>
          </string-name>
          .
          <article-title>The Architecture of the Roman Triumph: Monuments, Memory, and Identity</article-title>
          . Cambridge University Press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Reid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Savinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Teplyashin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lepikhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lillicrap</surname>
          </string-name>
          , J.-b. Alayrac,
          <string-name>
            <given-names>R.</given-names>
            <surname>Soricut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lazaridou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Firat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schrittwieser</surname>
          </string-name>
          , et al. “
          <article-title>Gemini 1.5: Unlocking Multimodal Understanding Across Millions of Tokens of Context”</article-title>
          .
          <source>Ianr:Xiv preprint arXiv:2403.05530</source>
          (
          <year>2024</year>
          ). doi: https://doi.org/10.48550/arXiv.2403.0553.0
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Schipporeit</surname>
          </string-name>
          . “
          <article-title>Wege des Triumphes. Zum Verlauf der Triumphzüge im spätrepublikanischen und augusteischen Rom”</article-title>
          . In:
          <article-title>Triplici invectus triumpho : Der römische Triumph in augusteischer Zeit</article-title>
          .
          <article-title>(</article-title>
          <year>2008</year>
          ), pp.
          <fpage>95</fpage>
          -
          <lpage>136</lpage>
          . url: https://zenon.dainst.org/Record/001 069375.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>N.</given-names>
            <surname>Stiennon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ouyang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ziegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Lowe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Voss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Christiano</surname>
          </string-name>
          . “
          <article-title>Learning to Summarize with Human Feedback”</article-title>
          .
          <source>InA:dvances in Neural Information Processing Systems</source>
          . Ed. by
          <string-name>
            <given-names>H.</given-names>
            <surname>Larochelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ranzato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Balcan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          . Vol.
          <volume>33</volume>
          . Curran Associates, Inc.,
          <year>2020</year>
          , pp.
          <fpage>3008</fpage>
          -
          <lpage>3021</lpage>
          . urlh:ttps://proceedin gs.neurips.cc/paper%5C%5Ffiles/paper/2020/file/1f89885d556929e98d3ef9b86448f951-Pa per.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>P.</given-names>
            <surname>Tennant</surname>
          </string-name>
          . “
          <article-title>The Lupercalia and the Romulus and Remus Legend”</article-title>
          .
          <source>InA:cta Classica</source>
          <volume>31</volume>
          (
          <year>1988</year>
          ), pp.
          <fpage>81</fpage>
          -
          <lpage>93</lpage>
          . url: http://www.jstor.org/stable/2459184.7
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>L.</given-names>
            <surname>Webb</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Brännstedt</surname>
          </string-name>
          . “
          <article-title>Gendering the Roman Triumph: Elite Women and the Triumph in the Republic and Early Empire”</article-title>
          .
          <source>InG:endering Roman Imperialism. Leiden, The Netherlands: Brill</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>58</fpage>
          -
          <lpage>95</lpage>
          .
          <year>do1i</year>
          :
          <fpage>0</fpage>
          .1163/9789004524774\_
          <fpage>005</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>B. L.</given-names>
            <surname>Welch</surname>
          </string-name>
          . “
          <article-title>The Generalization of Student's Problem when Several Diferent Population Variances are Involved”</article-title>
          .
          <source>InB:iometrika 34</source>
          .1-
          <fpage>2</fpage>
          (
          <year>1947</year>
          ), pp.
          <fpage>28</fpage>
          -
          <lpage>35</lpage>
          . doi:
          <volume>10</volume>
          .1093/biomet/34 .1-
          <fpage>2</fpage>
          .28. url: https://doi.org/10.1093/biomet/34.1-
          <fpage>2</fpage>
          .
          <fpage>28</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [32]
          <string-name>
            <surname>K</surname>
          </string-name>
          .-W. Welwei. “
          <article-title>Das Angebot des Diadems an Caesar und das Luperkalienproblem”</article-title>
          .
          <source>In: Historia: Zeitschrift für Alte Geschichte 16.1</source>
          (
          <issue>1967</issue>
          ), pp.
          <fpage>44</fpage>
          -
          <lpage>69</lpage>
          . url: http://www.jstor.org /stable/4434966.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dong</surname>
          </string-name>
          . “
          <article-title>ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation”</article-title>
          .
          <source>InA:dvances in Neural Information Processing Systems</source>
          . Ed. by
          <string-name>
            <given-names>A.</given-names>
            <surname>Oh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Globerson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Saenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Levine</surname>
          </string-name>
          . Vol.
          <volume>36</volume>
          . Curran Associates, Inc.,
          <year>2023</year>
          , pp.
          <fpage>15903</fpage>
          -
          <lpage>15935</lpage>
          . url: https://proceedings.neurips.cc/paper%5C%5Ffiles/paper/2023/file/33646ef0ed55414 5eab65f6250fab0c9-Paper-Conference.pd.f
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>