<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>If the Sources Could Talk: Evaluating Large Language Models for Research Assistance in History</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giselle Gonzalez Garcia</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christian Weilbach</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of British Columbia</institution>
          ,
          <addr-line>Vancouver</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of History and School of Irish Studies, Concordia University</institution>
          ,
          <addr-line>Montreal</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <fpage>616</fpage>
      <lpage>638</lpage>
      <abstract>
        <p>The recent advent of powerful Large-Language Models (LLM) provides a new conversational form of inquiry into historical memory (or, training data, in this case). We show that by augmenting such LLMs with vector embeddings from highly specialized academic sources, a conversational methodology can be made accessible to historians and other researchers in the Humanities. Concretely, we evaluate and demonstrate how LLMs have the ability of assisting researchers while they examine a customized corpora of different types of documents, including, but not exclusive to: (1). primary sources, (2). secondary sources written by experts, and (3). the combination of these two. Compared to established search interfaces for digital catalogues, such as metadata and full-text search, we evaluate the richer conversational style of LLMs on the performance of two main types of tasks: (1). question-answering, and (2). extraction and organization of data. We demonstrate that LLMs semantic retrieval and reasoning abilities on problem-specific tasks can be applied to large textual archives that have not been part of the its training data. Therefore, LLMs can be augmented with sources relevant to specific research projects, and can be queried privately by researchers.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Artificial Intelligence (AI)</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Large Language Models (LLMs)</kwd>
        <kwd>GPT</kwd>
        <kwd>Historical Research Methods</kwd>
        <kwd>Historical Writing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction &amp; Related Works</title>
      <p>
        Researchers in History and the Humanities tend to accumulate thousands of papers,
books, and other sources to be read, i.e. to be processed, in the near future [
        <xref ref-type="bibr" rid="ref24 ref34 ref5">5</xref>
        ]. However,
only a handful of these end up in the bibliographical section of our papers, dissertations,
and monographs. Because, we have to read them, i.e. process the information one page
at the time. Yet, we still personally archive all these sources in our physical or digital
libraries. Departing from this premise, we propose: what if these collections of academic
texts can be incorporated into a corpus, infused into a Large Language Model (LLM),
such as ChatGPT, and interrogated in new ways to produce better informed research
outputs? This informed dialogue between researcher and machine would not necessarily
be an output in itself, but it would significantly accelerate the research process in History
and the Humanities.
      </p>
      <p>
        LLMs can be defined as a computer system that, given a word or text (token), can
predict the words that would come after. Although these systems are not infallible,
many researchers conceptualize the advent of LLMs as a proposition of full automation,
especially in creative fields. In referring to full automation, we allude to the increasingly
popular believe that LLMs entail complete replacement of humans in the production of
knowledge and historical narratives. But considering the history of technology since the
Industrial Revolution, machines have automated production processes that were often
repetitive and exhausting for human labour. Automation freed human resources, it has
provided humans with agency, and given us the possibility of dedicating ourselves to
higher cognitive tasks [
        <xref ref-type="bibr" rid="ref26 ref36 ref7">7</xref>
        ].
      </p>
      <p>In the field of History, little attention has been devoted to understanding how using
LLMs as part of our daily research praxis can have a deep impact and be a methodological
game changer. LLMs have the potential to modify the way traditional historical archives
are perused, the way primary sources are read and processed, the way theories and
narratives are probed and validated, the way complex social processes are summarised,
and finally, the way histories themselves are written. Our paper stems from the desire of
incorporating LLMs into the authors’ professional praxis in History and the Humanities.
In doing so, we do not expect the LLMs to "tell the truth" for us, or to produce their own
historical narrative in the form of synthetic new texts. Our main goal is to demonstrate
what areas of the historical research process can be significantly enhanced and automated
by having access to high-quality and accurate assistance. Even though, our professional
experience is within the fields of History and Machine Learning, many of the case studies
here exemplified can also appeal to other disciplines across the Humanities, and can
illuminate the path forward in the field of Artificial Intelligence.</p>
      <p>
        When addressing the emergence of LLMs, several areas of concern are mentioned:
(1). its potential abuse by students to plagiarize class assignments, (2). its incapacity
to distinguish truth from falsehood [
        <xref ref-type="bibr" rid="ref10 ref29">10</xref>
        ], (3). its environmental impact [
        <xref ref-type="bibr" rid="ref2 ref21 ref31">2</xref>
        ], and (4) its
ethical biases and risks [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. In the words of Wulf Kansteiner, "Large language models
such as GPT-3 are able to generate compelling, non-plagiarized texts in response to
simple natural language inputs, thus providing students with an opportunity to produce
high-quality written assignments with minimum effort." [
        <xref ref-type="bibr" rid="ref10 ref29">10</xref>
        ]. One of the main limitations
of LLMs (specifically of GPT-3), that Kansteiner identified, is its structurally inability
"to attribute the statements it generates to a specific textual origin, let alone assess
the factual reliability of any of its textual inputs or outputs." [
        <xref ref-type="bibr" rid="ref10 ref29">10</xref>
        ]. This is a problem
to which our paper offers a tentative solution by arguing that combining an LLM with
carefully selected corpora made of primary and secondary sources reduces the margins
for errors, and produces texts that have an origin in peer-reviewed sources. A similar
approach has been already implemented by Manjavacas Arevalo and Fonteyn [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] to
positive results in the form of the MacBERTh model. But, unlike them, we do not train
a new model, but augment pre-trained ones.
      </p>
      <p>
        Moreover, the answers we intend on retrieving from the LLMs include specific details
about its sources in the form of text chunks. Therefore, instead of getting answers to our
questions from models pre-trained on data only acquired from the world wide web, we
propose to prompt the models with a context of chunks that have already been through
the established process of academic validation, which can also be checked to validate the
response. This approach also shows promise to train better LLMs in general [
        <xref ref-type="bibr" rid="ref27 ref37 ref8">8</xref>
        ].
      </p>
      <p>
        Bender, Gebru, McMillan-Major, and Shmitchell [
        <xref ref-type="bibr" rid="ref2 ref21 ref31">2</xref>
        ], present a critical overview of the
most problematic aspects of LLMs, including their significant contributions to climate
change, and their tendency to reproduce and validate hegemonic ideas that are
overrepresented on the internet, and therefore on the training data that infuses these models.
At the same time LLMs show promise to scale deliberation and integrate diverse view
points in democratic processes [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. We do not intend to add to this debate at this early
stage in our research, but we acknowledge the current academic debate that
problematizes the multiple societal paradoxes that the advent of LLMs (ChatGPT, in particular)
have brought to the forefront.
      </p>
      <p>
        To evaluate such an scenario of historical research assistance empirically, we
implemented an open-source pipeline (Figure 1) with different freely available LLMs that
can be run locally (and privately) by any researcher (Section 3). We call the
implementation KleioGPT. We then compiled a comprehensive academic corpus of digitized
history monographs from the first author’s academic research in Irish Migration Studies,
and conducted a case study from the perspective of a professional historian in Section 4.
Our case studies consist of two important textual research tasks, (1). question-answering
in Section 4.1 and (2). data extraction in Section 4.2. We assembled 40 prompts for the
question-answering task and graded the answers given by different LLMs. We took into
account whether the LLMs really responded to our prompt correctly, and paid
attention to which sources from our corpus were pulled for answering correctly. For testing
data extraction we resorted to the 9 volume genealogical collection Historia de Familias
Cubanas by Santa Cruz y Mallen [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>Since the texts from the different corpora are fed to the LLMs in their general textual
representation, any other form of document retrieval from an archive, such as full text,
web or metadata search would equally work. We have not explored this in this work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>Our pipeline shown in Figure 1 consists of a two stage process. In the first stage, the
textual content of all documents in the archive is extracted and sliced into equally-sized
overlapping chunks. These chunks are then individually stored into a vector database
from which they can be efficiently retrieved later. To answer a question, first a number
of relevant chunks are queried from the vector database, and then optionally fed to
the LLM together with the question as its prompt. This allows the LLM to answer
archive-specific questions that exceed the information of its training data.</p>
      <sec id="sec-2-1">
        <title>2.1. Vector databases</title>
        <p>Vector databases index so called “embeddings”. These embeddings summarize the
content of a text chunk from a document in a fixed length vector of real numbers. To create
the database, first each chunk is paired with its respective embedding. The chunk is
then added under its embedding to the index of the database. Once embedded, these
vectors have the property of locating semantically similar documents near each other.
While executing a query, the database refers to its index to find those documents located
in proximity to the embedding that matches the text of the query. It then retrieves the
text chunk of each nearby embedding from the archive.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Large language models</title>
        <p>
          Large language models are a form of generative model. Generative models are
probabilistic models that can be fit to a data set and generate synthetic data that mimics it.
In our work, we depart from the successful class of generative pre-trained transformer
(GPT) models [
          <xref ref-type="bibr" rid="ref18 ref23 ref33 ref4">4, 18</xref>
          ]. GPT models are pre-trained on large public and private archives
to predict the next word in a given context. This type of pre-training has proven to be
effective in yielding state-of-the-art language models that perform well when given tasks
of interest for researchers in the Humanities [
          <xref ref-type="bibr" rid="ref12 ref28 ref38 ref9">9, 12</xref>
          ]. These include question-answering
and summarization, a comprehensive recent survey is provided in Yang, Jin, Tang, Han,
Feng, Jiang, Yin, and Hu [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
        <p>While pre-training such a model on billions of words is very costly, both in terms of
time and compute, running these pre-trained models is possible on commodity hardware.
For our work, we selected competitive (at the time of their release) publicly downloadable
models, as well as the most popular models, such as GPT3 and ChatGPT; see Table 1.
For the open models we chose small versions that can be run on consumer graphics
cards (Appendix A). The primary interface to LLMs is the so called prompt, which is
the text provided to the model to describe the context and task at hand. The model then
appends its answer to the prompt as a form of auto-completion by next word prediction.
We document our prompts for each experiment in Appendix B. Generation in LLMs is
generally stochastic and the amount of stochasticity can be modulated by a so called
temperature. A common problem in LLMs is so called hallucination (see Appendix C),
which refers to the fact that they can make up answers that have no factual basis.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>We integrated available LLM open-source software to execute the task described in
Section 3.1 together with an interactive question-answering mode in a Python code
base for KleioGPT.1 For Section 4.2 we directly prompt ChatGPT as described there.</p>
      <sec id="sec-3-1">
        <title>3.1. Retrieval augmented question-answering</title>
        <p>
          Our methodology is based on the established combination of memory retrieval from
vector databases with LLMs [
          <xref ref-type="bibr" rid="ref22 ref3 ref32">3</xref>
          ]. Augmenting sources does not only help the LLM to
answer the question, it also provides a way to check whether the answer can be backed
up by the documents in the vector database, a step we think is necessary in an academic
setting. Our implementation uses a pre-trained transformer embedding model,2 the
vector database Chroma,3 and, to integrate the different LLMs, we use LangChain.4
For a fair and reproducible comparison we set the temperature of all models (except for
ChatGPT where we have no control) close to zero (10−5), rendering generation effectively
deterministic. A zero temperature corresponds to picking the most likely answer in each
step and encourages factual correctness over creativity. We chose the default settings of
each model in LangChain for all other parameters. We improved the retrieval mechanism
of LangChain to filter out chunks of text that are from bibliographic sections of the
sources and chunks that contain less than 200 characters. During question-answering,
we loaded a set of questions from a prepared comma separated value (CSV) file. In return,
we obtained another CSV file as a report with the initial questions, the answers, and the
retrieved sources. Examples are shown in Section 4.1. This simple process both lends
itself to automated batch-processing of questions and is accessible to non-technicians
through its tabularized inputs and outputs.
1https://github.com/GissyGonzalez/KleioGPT
2https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2
3https://docs.trychroma.com/
4https://langchain.com/
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Datasets</title>
        <sec id="sec-3-2-1">
          <title>3.2.1. Migration Studies Corpus</title>
          <p>To examine the question-answering capacity of each LLM, we assembled a corpus that
forms our Migration Studies dataset. Our aim is to show that researchers can assemble
their own corpora tailored to their specific projects. This bibliographic corpus is made
of 86 books from leading scholars in Irish Migration Studies, Cuban Studies, general
Migration Theory, histories of Slavery, etc. They have been selected from the main
fields of study of the first author, and are the basis for their doctoral research. This
corpus is representative of the best works in these fields, and among its authors are the
leading academic experts in each subject.</p>
          <p>In order to verify (and grade) the answers provided by each LLM, we selected volumes
the first author was familiar with, had read recently, and had a knowledge of each
writer’s placement within the historiography and the broader academic conversation in
themes that include migration, race, gender, and slavery. These are mostly historical
monographs that have been through a peer-review publication process. This does not
exempt them from reproducing the societal biases of their authors, but we believe no
text is exempt of reproducing the ideologies and cultural systems researchers subscribe
to. Nonetheless, we prioritized feeding KleioGPT sources that have gone through a
wellestablished academic validation process, and can be generally factually trustworthy. We
provide a list of these sources in Appendix D.</p>
          <p>As mentioned above, researchers can gather their own corpus and ingest them into
their own versions of KleioGPT. Every text loaded into our experiments was in PDF
format. We made sure that each PDF had a plain text layer that ensured its content
is machine readable. Users should make sure that their OCR’ed texts are as clean as
possible for better results. Our Migration Studies corpus is exclusively in the English
language, although this methodology can be applied to corpora in other languages if the
embeddings and LLMs support it.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Family History Books Compilation</title>
          <p>In the past five years a considerable part of the first author’s doctoral research on Irish
migrants in the Hispanic Caribbean has consisted of reviewing genealogical compilations
in search for biographical data related to 19th-century migrants. In many occasions,
genealogical data is the only type of information retrievable to rescue the histories of
this population.</p>
          <p>Genealogical sources are usually structured in a regular pattern that outlines birth,
marriage, death facts, and kinship connections among people. These are worded similarly.
Automating the extraction of this type of data from family history book compilations is
a task that many historians and researchers in the Humanities undertake frequently. It
is labour intensive, repetitive, and consumes significant amounts of time that could be
devoted to other research endeavours. Therefore, the possibility of automation by using
LLMs is a very attractive one.</p>
          <p>
            Our genealogical dataset is made of the 9 volumes of Historia de Familias Cubanas
(a) Word cloud of the 9 volumes of Histo- (b) Links visualization of the 9 volumes of
ria de Familias Cubanas by Francisco Historia de Familias Cubanas by
FranJavier de Santa Cruz [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] showing the cisco Javier de Santa Cruz showing the
word frequency and the regularity of regular structure of genealogical
lannames and expressions. Made by the guage corpus. Made by the authors
authors with VoyantTools. with VoyantTools.
          </p>
          <p>
            Santa Cruz y Mallen [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] (including OCR). This corpus consists of 1,803,596 total words
and 39,114 unique word forms. The data set is visualized in Section 3.2.2.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <sec id="sec-4-1">
        <title>4.1. Academic question-answering</title>
        <p>To interrogate our corpus from Section 4.1 we designed a fixed set of questions and
followed the methodology outlined in Section 3.1. Our goal was to examine and
compare the quality of different LLMs responses to four different types of queries: factual,
argumentative, descriptive, and integrative. Their level of complexity corresponded to
the types of questions undergraduate student in the 200 and 300 level courses would
typically face. We graded the answers in a pass/failed binary, giving one point for
correct answers, and zero for the opposite. We also knew beforehand the texts contained
accurate answers to these questions. All the questions can be found in Appendix B.</p>
        <p>We summarize our results in Table 2. All LLMs could answer the majority of questions
in all configurations. The retrieved text chunks from the vector database in general
improve the results and improve important context to validate the answer, but the
LLMs are interestingly already able to answer many questions even without any sources.
ChatGPT performed best even without any sources and remarkably we found that there
is a big gap to GPT3 accessed through the OpenAI API, which is underlying ChatGPT.
There are probably significant extensions applied to the web interface of ChatGPT.
Both XGen and Beluga consistently outperform GPT3 and both are competitive with
ChatGPT.</p>
        <p>
          Acting under 0 chunks, i.e. without having access to our corpus, the LLMs tended
to fail when giving answers to specific questions like: "What was the population of
Ireland in 1841?". They were also not able to summarize the contents and main ideas
of books like Kerby A. Miller’s Emigrants and Exiles [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Without the sources in our
dataset, LLMs like Falcon were still able to define conceptual ideas and answer correctly
prompts like: "Can you define what is migration/Irish nationalism?". Overall, tasks
that required the LLMs to generate its own ideas or argumentation (like those in the
integrative section of our Q&amp;A) tended to perform worse, or even prompted the LLMs
to hallucinate (see Appendix C) completely off-the-mark answers, when it did not have
the backup of the corpus.
        </p>
        <p>Factual The ten questions dedicated to retrieval of factual information focused on
obtaining precise data such as dates, quantities, locations, etc. Fact-checking queries
are very typical during the historical research process. When these types of question are
posed to search engines like: Google, Bing, Startpage, etc., these point us to a list of
often millions of indexed results. We have to open web pages to retrieve the information
we are looking for and evaluate it. As a system to effectively retrieve factual data,
LLMs give us a direct answer, therefore shortening significantly the time invested on
this task. These types of questions were easy to grade because they were either right or
wrong. When some answers were an approximate, results were graded according to the
standard academic consensus.</p>
        <p>Question 5: "Where in Ireland are Protestants found?"
XGen: "Protestants were found in all parts of Ireland, but they were
particularly concentrated in urban areas and in the southern province of Ireland.
In the early 1800s, they were a majority in County Londonderry."
Falcon: "Protestants are found in all parts of Ireland, but are more commonly
found in urban areas."
GPT3: "Protestants are found throughout Ireland, but they tend to be
concentrated in towns and urban areas."
ChatGPT: "Protestants in Ireland can be found throughout the country, but
their distribution varies across different regions. Historically, Protestants
have had a stronger presence in Northern Ireland, particularly in the
counties of Antrim, Down, Armagh, and Londonderry (also known as Derry). In
these areas, Protestant communities, including various denominations such
as Anglicans, Presbyterians, and Methodists, have traditionally been more
prominent. However, it’s important to note that there are Protestant
communities and individuals living in all parts of Ireland, including the Republic
of Ireland."</p>
        <p>All models answer Question 5 correctly, but ChatGPT provides a lot more relevant
details. It is unclear why it behaves so differently, but it could be that it is provided an
additional prompt that is not publicly accessible.</p>
        <p>Argumentative We also designed ten questions for this type of query. In this category,
we asked for summaries of concepts, ideas, and entire books. It is important to clarify
that, once our corpus was added, LLMs summarize books based on other scholars
opinion’s of those books retrieved from the chunks, and not on their own complete reading of
each book.5 Essentially, it will summarize what other authors said about a specific book,
and not the book itself. This can be valuable to realize what the academic consensus is
about a specific work, idea, or theory; and to identify contentious arguments about them.
Most questions in this category that the LLMs failed, were because they hallucinated or
fabricated events that were easily identifiable as factually wrong.</p>
        <p>Question 3: "Can you summarize the three main arguments in Kerby Millers
book Emigrants and Exiles?"
XGen: "1. Irish American nationalism is the primary expression of a powerful
sense of exile among Irish immigrants to America. 2. Irish immigrants to
America were deeply homesick and resistant to the culture of individualism.
3. Irish chain migration to North America was not passive, but rather relied
on the recommendations of previous migrants."
GPT3: "I don’t know the answer."
ChatGPT: "Kerby Miller’s book "Emigrants and Exiles: Ireland and the
Irish Exodus to North America" explores the causes, experiences, and
consequences of Irish emigration to North America during the 18th and 19th
centuries. While it is challenging to fully summarize all the arguments in the
book, here are three key themes..."
5The recent release of Claude 2 (https://claude.ai) with a large context window of 100.000 tokens
provides such functionality, unfortunately it was not yet programmatically available at the time of
submission to the authors.</p>
        <p>In the example above, the authors were expecting a succinct answer similar to the
one provided by XGen. GPT3’s reply was unexpected given that Miller’s book is a
classic volume in the field. Emigrants and Exiles (1985) has been discussed in length on
the web, and according to Google Scholar, is cited 1663 times. Discussions about this
book should be on GPT3 training data. Its response contrasts starkly with ChatGPT’s
lengthy and detailed one.</p>
        <p>Descriptive Raising the level of complexity, in this category, we included questions
that necessitated more elaborated answers about historical events and processes.</p>
        <p>Question 4: "Can you describe the events that led to the Great Irish potato
Famine?"
ChatGPT: "The Great Irish Potato Famine, also known as the Great Hunger
or the Irish Potato Famine, was a devastating period in Irish history that
occurred between 1845 and 1852. It was a catastrophic event primarily caused
by a combination of potato crop failure, economic policies, and socio-political
factors. Here is a description of the events that led to the famine: Potato
Dependence: ..."</p>
        <p>All models answered this question correctly. This can be attributed to how pervasive
in society the topic at hand is and how much it has been discussed online. XGen again
gave a correct concise answer, however we were expecting a lengthier description given
the complexity and gravity of the Great Famine. ChatGPT not only gave a very detailed
answer, but it also outlined the causal chain of events, step by step, leading to the famine
and its extent (the full answer is in Appendix B.2). Unlike ChatGPT, the other LLMs
provided answers that were below the standard of what would have been expected of an
early undergraduate Irish Studies student.</p>
        <p>Integrative To elaborate this type of questions, we took inspiration in the type of
exercises graduate students in North America are tested on during their comprehensive
examinations. In answering these questions, the LLMs needed to demonstrate more
creativity than in the previous ones. This section was supposed to entail the highest
level of difficulty to be tested during our experiments. Here, the LLMs struggled with
replying in accordance to the level expected of a human graduate student in History.</p>
        <p>Question 3: "What is the future of History as a discipline?"
XGen: "The future of history as a discipline is uncertain and likely to be
shaped by a variety of factors, including changes in technology, the rise of
new historical methodologies, and shifts in the broader cultural and political
landscape. Some experts predict that history will continue to evolve and
become more interdisciplinary, incorporating insights from other fields such as
anthropology, sociology, and economics. Others suggest that the rise of
digital humanities and the increasing availability of large datasets may lead to
new approaches to historical research and analysis. Additionally, the future
of history may be influenced by broader trends such as the increasing
importance of global and transnational history, as well as the growing interest in
the history of science and technology. Ultimately, the future of history will
depend on the choices and priorities of historians and the broader academic
community."
GPT3: "Based on the given context, it is not possible to accurately determine
the future of history as a discipline. The context provides information about
the challenges and changes that the discipline of history has faced, such as
the questioning of the nature of history, the interaction with humanities, arts,
and sciences, and the interdisciplinary nature of new scholarship. However,
there is no explicit information or discussion about the future trajectory of
the discipline."</p>
        <p>In the example provided, we observe the contrast between two different approaches:
one in which GPT3 determined it had not been provided enough material from the
sources and is therefore uncertain about how to answer the question; and another in
which XGen based on the same materials, tried to give a well-rounded reply.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Extracting Tabular Data from Genealogical Sources</title>
        <p>
          In this section, we took several pages from Historia de Familias Cubanas (Section 3.2.2)
and manually provided their full contents to the web interface of ChatGPT (the original
text can be seen in Figure 3). The content of these 9 volumes is entirely in Spanish. Even
though only 7% of GPT3 training data is in a language other than English [
          <xref ref-type="bibr" rid="ref23 ref33 ref4">4</xref>
          ], ChatGPT
performed well in Spanish, translating accurately the excerpts provided. Afterwards, we
designed the following prompt:
        </p>
        <p>Prompt example: "From the previous text, list all the names of people in
a table with columns: full name of each person, relationship, date of birth</p>
        <p>Full Name
Domingo de Ajuría
Isabel de Mendibíl
Francisco de Ajuria y Mendibil
Isabel Urratia y Gordobil
Tomás de Ajuria y Urratia
Elena Goiri e Irizarri
Francisco de Ajuria y Goiri</p>
        <p>Date of Birth
Unknown
Unknown
Unknown
Unknown
15-03-1671
Unknown
Unknown</p>
        <p>Place of Birth
Ubidea
Unknown
Ubidea
Unknown
Ubidea
Unknown
Unknown</p>
        <p>Baptism Date
Unknown
Unknown
Unknown
Unknown
15-03-1671
Unknown
Unknown</p>
        <p>Marriage Date
Unknown
Unknown
28-07-1664
28-07-1664
12-08-1693
12-08-1693
Unknown
Father’s Full Name
Unknown
Unknown
Domingo de Ajuría
Martin
Francisco de Ajuria
Domingo
Tomás de Ajuria</p>
        <p>Mother’s Full Name
Unknown
Unknown
Isabel de Mendibíl
Ana
Isabel Urratia
Maria
Elena Goiri e Irizarri</p>
        <p>Children’s Full Name
Francisco de Ajuria y Mendibil
Unknown</p>
        <sec id="sec-4-2-1">
          <title>Tomás de Ajuria; Francisco de Ajuria</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Tomás de Ajuria; Francisco de Ajuria</title>
          <p>Francisco de Ajuria y Goiri
Francisco de Ajuria y Goiri</p>
          <p>Unknown
Spouse’s Full Name
Isabel de Mendibil
Domingo de Ajuría
Isabel Urratia y Gordobil
Francisco de Ajuria
Elena Goiri e Irizarri
Tomás de Ajuria y Urratia
Unknown</p>
          <p>Gender
Male
Female
Male
Female
Male
Female
Male
(format DD-MM-YYYY), place of birth, date of death (format
DD-MMYYYY), baptism date (format DD-MM-YYYY), marriage date (format
DDMM-YYYY), place of residence, full name and surname of father, full name
and surname of mother, full name and surname of children, full name and
surname of spouse, and occupation. Try to infer the gender of each person,
and add a column Gender."</p>
          <p>
            In Table 3 we show that ChatGPT is able to extract data from Historia de Familias
Cubanas (9 volumes) in Santa Cruz y Mallen [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] into a structured tabular format by
simply prompting it to create a table with the respective column names.
          </p>
          <p>A large part of the information is correctly mapped into the table. However, ChatGPT
did miss some of the information highlighted in bold. The missing fragments show
regularities such as family names that are connected by the conjunction “y”. We believe
that it is likely that by instructing ChatGPT explicitly on these expressions, e.g. by
giving example translations or erroneous lines, the LLM would miss less data. We plan
to address this in future work together with an automatic page wise extraction of the
tables with our prompt.</p>
          <p>As can be seen in Figure 3, the gender information was not in the original text.
We asked ChatGPT to infer it automatically based on its pre-trained knowledge about
Spanish name conventions.</p>
          <p>Out of all the dates extracted, there is only one birth date (in red) inferred wrongly,
everything else is correct. If this low error rate reflects the performance on the full
dataset, then a large amount of data could be extracted into a tabular representation
just by asking ChatGPT to do so. We could not get the same extraction quality out of
the other LLMs, but expect them to catch up in the near future.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper we have demonstrated that LLMs can provide a fluent conversational
research assistance while being sufficiently accurate for an academic environment. While
open source models that fit onto commodity hardware are not yet competitive with
ChatGPT in all tasks, they already provide a private alternative on question-answering
tasks on sensitive data. While this work already highlights the potential of LLMs for
historical research, problems such as hallucination and biases in the training data will
also probably be significantly reduced in the near future. Since the field of LLMs is
evolving very quickly, we think that very soon much better open source LLMs will be
available. We will continue to design tests to evaluate the LLMs abilities in the near
future.</p>
      <p>Interesting avenues for future work are better integration of memory to both help
the generation of the LLMs and make it easier to validate them against the sources
retrieved from memory. Models such as Claude 2 make use of larger context windows for
better extraction and summarization abilities and show a lot of promise in our ongoing
evaluations as well. We think that open source solutions that facilitate research will be
widely available to researchers in the Humanities including models that can be run on
machines without GPUs.6
6An older example for this is https://gpt4all.io/index.html</p>
    </sec>
    <sec id="sec-6">
      <title>A. Experimental details</title>
      <p>We ran experiments for the local LLMs StableBeluga-7b, falcon-7b-instruct and
xgen7b-8k-inst on an A5000 NVIDIA GPU.</p>
    </sec>
    <sec id="sec-7">
      <title>B. Academic question-answering</title>
      <sec id="sec-7-1">
        <title>B.1. Questions</title>
        <p>B.1.1. Factual
B.1.3. Descriptive
1. Can you describe Irelands population loss process throughout the nineteenth
century?
2. What was the Young Ireland movement?
3. What was the Land League?
4. Can you describe the events that led to the Great Irish potato Famine?
5. What were the main five consequences for Ireland of the Potato Famine of 1845?
6. What was the Catholic Emancipation movement?
7. Can you describe how was the Irish Act of Union passed in Parliament?
8. Can you describe chain migration?
9. What was the political state of Ireland in 1845?
10. Who was Daniel OConnell?
B.1.4. Integrative
1. What are the two most significant controversies that have developed in Irish
diaspora studies in the past twenty-five years? Elaborate.
2. What areas of study should Irish history researchers focus on in the future?
Elaborate.
3. What is the future of history as a discipline? Elaborate.
4. Which geographic areas and historical time periods do you think should receive
more or less attention based on what are the dominant themes in the current Irish
Historiography? Elaborate.
5. What elements do you consider to be crucial in teaching Irish diaspora studies, and
what do you think needs to be revised? Elaborate.
6. What is the future of migration studies? Elaborate.
7. Taking from Kerby Millers Emigrants and Exiles book, and for Donald Harman
Akensons book The Irish in Ontario, what are the main arguments these
historian make, how are these arguments similar and different, and what are contested
subjects in their works? Elaborate.
8. How does Irish migration to the United States differ from Irish migration to</p>
        <p>Canada? Elaborate.</p>
        <p>9. Were the Irish in the United States considered white? Elaborate.
10. What were Irish migrants role in African slavery? Elaborate.</p>
      </sec>
      <sec id="sec-7-2">
        <title>B.2. ChatGPT answer to Irish Potatoe Famine</title>
        <p>ChatGPT: "The Great Irish Potato Famine, also known as the Great Hunger or the
Irish Potato Famine, was a devastating period in Irish history that occurred between
1845 and 1852. It was a catastrophic event primarily caused by a combination of potato
crop failure, economic policies, and socio-political factors. Here is a description of the
events that led to the famine: Potato Dependence: The Irish population in the mid-19th
century heavily relied on the potato as a staple food crop. The potato was affordable,
easy to grow, and provided sufficient calories to sustain a large population. However,
this heavy dependence on a single crop made the population vulnerable to any failures
or diseases affecting the potato harvest. Potato Blight: In 1845, a strain of potato
blight, caused by a fungal disease called Phytophthora infestans, spread across Ireland.
The blight rapidly destroyed the potato crops, turning them into a black, rotting mass.
This initial outbreak resulted in a significant reduction in the potato harvest, leading
to food shortages and increasing fears of famine. British Economic Policies: Ireland
was under British rule during this time, and British economic policies exacerbated the
impact of the blight. British policies favored exporting Irish agricultural produce, such
as grain and livestock, to Britain while leaving the Irish population heavily dependent
on the potato as their main food source. The policies prioritized the interests of British
landowners and exacerbated existing economic inequalities in Ireland. Limited Relief
Efforts: The British government’s response to the crisis was initially insufficient. Relief
efforts were limited, and policies were implemented with a focus on maintaining economic
stability rather than prioritizing the well-being of the Irish population. This lack of
effective intervention worsened the situation and increased the suffering of those affected
by the famine. Rapid Spread and Devastation: The potato blight continued to devastate
subsequent potato crops in the years that followed. The impact was particularly severe
in the west and southwest regions of Ireland, where the population was heavily reliant
on potato cultivation. The loss of the potato harvest resulted in widespread hunger,
malnutrition, and disease. Mass Emigration and Death: The combination of crop failure,
lack of relief, and dire living conditions led to mass emigration and significant loss of life.
Many Irish people, desperate to escape the famine’s effects, left their homes and families
to seek opportunities in other countries, particularly the United States and Canada.
However, emigration was not an option for everyone, and the lack of adequate food and
resources resulted in a devastating death toll, estimated to be around one million people,
while millions more emigrated. The Great Irish Potato Famine had a profound and
lasting impact on Ireland. It led to depopulation, cultural trauma, and a reevaluation of
Irish-British relations. The famine exposed the socio-economic inequalities, poor relief
systems, and inadequate governance during that time, and its effects continue to be
remembered as a defining period in Irish history."</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>C. Hallucination Example</title>
      <p>Question 9 (Argumentative): "Can you summarize what David Fitzpatricks
book Oceans of Consolation is about?"
GPT3: "David Fitzpatrick’s book "Oceans of Consolation" is a fundamental
study and model for further editions of emigrant letters. It explores the
personal letters sent to and from Irish migrants in Australia, providing full
transcripts and examining each correspondent’s local context. The book aims
to understand the settlement of the Irish abroad in a comparative context
and emphasizes the importance of faith in the lives of Irish migrant."
Falcon: "Ocean of Consolation is a book about the history of the ocean
and its impact on human culture. It explores the various ways the ocean has
been perceived throughout history, from being seen as a source of wealth and
power to being a place of spiritual significance. The book also examines the
current state of the world’s oceans and the threats they face due to human
activity."</p>
    </sec>
    <sec id="sec-9">
      <title>D. Irish Migration Studies Dataset</title>
      <p>Stradling, R. A. 1994. The Spanish Monarchy and Irish Mercenaries : The Wild Geese
in Spain 1618-68. Blackrock Co. Dublin: Irish Academic Press.</p>
      <p>Valente, Joseph. 2011. The Myth of Manliness in Irish National Culture 1880-1922.
Urbana Ill: University of Illinois Press.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>E.</given-names>
            <surname>Almazrouei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Alobeidli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Alshamsi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cappelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cojocaru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Debbah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Goffinet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Heslow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Launay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Malartic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Noune</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pannier</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Penedo.</surname>
          </string-name>
          “
          <article-title>Falcon-40B: an open large language model with state-of-the-art performance”</article-title>
          . In: (
          <year>2023</year>
          ). URL: https://falconllm.tii.ae/.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Bender</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gebru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>McMillan-Major</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Shmitchell</surname>
          </string-name>
          . “
          <article-title>On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?”</article-title>
          <source>In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. Virtual Event Canada: Acm</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>610</fpage>
          -
          <lpage>623</lpage>
          . DOI:
          <volume>10</volume>
          .1145/3442188.3445922.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Borgeaud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hoffmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Rutherford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Millican</surname>
          </string-name>
          , G. van den Driessche, J.
          <string-name>
            <surname>-B. Lespiau</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Damoc</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
          </string-name>
          , D. d. L.
          <string-name>
            <surname>Casas</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Guy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Menick</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Ring</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Hennigan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Maggiore</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Cassirer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Brock</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Paganini</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Irving</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Vinyals</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Osindero</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Simonyan</surname>
            ,
            <given-names>J. W.</given-names>
          </string-name>
          <string-name>
            <surname>Rae</surname>
            , E. Elsen, and
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Sifre</surname>
          </string-name>
          .
          <source>Improving Language Models by Retrieving from Trillions of Tokens</source>
          .
          <year>2022</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2112.04426. arXiv:
          <volume>2112</volume>
          .04426 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            , E. Sigler,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Litwin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Chess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Berner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McCandlish</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>I. Sutskever</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          . “
          <article-title>Language Models Are Few-Shot Learners”</article-title>
          . In: arXiv:
          <year>2005</year>
          .14165 [cs] (
          <year>2020</year>
          ). arXiv:
          <year>2005</year>
          .14165 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Crane.</surname>
          </string-name>
          “
          <article-title>"What Do You Do with a Million Books?"”</article-title>
          .
          <source>In: D-Lib Magazine 12.3</source>
          (
          <year>2006</year>
          ). URL: https://www.dlib.org/dlib/march06/crane/03crane.html.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Erik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hayashi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xing</surname>
          </string-name>
          , S. Vig Jesse ang Yavuz,
          <string-name>
            <given-names>P.</given-names>
            <surname>Laban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Krause</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Purushwalkam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Kryscinski</surname>
          </string-name>
          , L. Murakhovs'ka,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Choubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fabbri</surname>
          </string-name>
          , Y. Liu,
          <string-name>
            <given-names>R.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bhat</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-S. Wu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Savarese</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>S. R.</given-names>
          </string-name>
          <string-name>
            <surname>Joty</surname>
          </string-name>
          , and
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Xiong. Long Sequence Modeling with XGen: A 7B LLM Trained on 8K Input Sequence Length</article-title>
          .
          <source>Salesforce AI Research Blog</source>
          .
          <year>2023</year>
          . URL: https://blog.salesforceairesearch.com/xgen.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C. B.</given-names>
            <surname>Frey</surname>
          </string-name>
          .
          <article-title>The Technology Trap Capital, Labor, and Power in the Age of Automation. First paperback printing</article-title>
          .
          <source>Book Collections on Project MUSE</source>
          . Oxford: Princeton University Press,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gunasekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aneja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C. T.</given-names>
            <surname>Mendes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Del Giorno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gopi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Javaheripi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kauffmann</surname>
          </string-name>
          , G. de Rosa,
          <string-name>
            <given-names>O.</given-names>
            <surname>Saarikivi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. S.</given-names>
            <surname>Behl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bubeck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Eldan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Kalai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. T.</given-names>
            <surname>Lee</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <source>Textbooks Are All You Need</source>
          .
          <year>2023</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2306.11644. arXiv:
          <volume>2306</volume>
          .11644 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Beelen</surname>
          </string-name>
          , G. Colavizza, and
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Ardanuy</surname>
          </string-name>
          . “
          <article-title>Neural Language Models for Nineteenth-Century English”</article-title>
          .
          <source>In: 7.0</source>
          (
          <issue>2021</issue>
          ), p.
          <fpage>22</fpage>
          . DOI:
          <volume>10</volume>
          .5334/johd.48w.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Kansteiner</surname>
          </string-name>
          . “
          <article-title>Digital Doping for Historians: Can History, Memory, And Historical Theory Be Rendered Artificially Intelligent?”</article-title>
          <source>In: History and Theory 61.4</source>
          (
          <issue>2022</issue>
          ), pp.
          <fpage>119</fpage>
          -
          <lpage>133</lpage>
          . DOI:
          <volume>10</volume>
          .1111/hith.12282.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Klein</surname>
          </string-name>
          . “
          <source>Are Large Language Models Our Limit Case?” In: Startwords</source>
          <volume>3</volume>
          (
          <year>2022</year>
          ). DOI:
          <volume>10</volume>
          .5281/zenodo.6567985.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>E. Manjavacas</given-names>
            <surname>Arevalo</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Fonteyn</surname>
          </string-name>
          . “
          <article-title>MacBERTh: Development and Evaluation of a Historically Pre-trained Language Model for English (1450-1950)”</article-title>
          .
          <source>In: Proceedings of the Workshop on Natural Language Processing for Digital Humanities. NIT Silchar</source>
          ,
          <source>India: NLP Association of India (NLPAI)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Emigrants and Exiles: Ireland and the Irish Exodus to North America</article-title>
          . New York; Oxford: Oxford University Press,
          <year>1985</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mitra</surname>
          </string-name>
          , G. Jawahar,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Palangi</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Awadallah</surname>
          </string-name>
          .
          <source>Orca: Progressive Learning from Complex Explanation Traces of GPT-4</source>
          .
          <year>2023</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2306.02707. arXiv:
          <volume>2306</volume>
          .02707 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>F. X. de Santa Cruz y Mallen. Historia De Familias Cubanas (Tomos</surname>
            <given-names>I-IX</given-names>
          </string-name>
          ).
          <year>1940</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Small</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Vendrov</surname>
          </string-name>
          , E. Durmus,
          <string-name>
            <given-names>H.</given-names>
            <surname>Homaei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Barry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cornebise</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Suzman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ganguli</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Megill</surname>
          </string-name>
          .
          <article-title>Opportunities and Risks of LLMs for Scalable Deliberation with Polis</article-title>
          .
          <year>2023</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2306.11932. arXiv:
          <volume>2306</volume>
          .11932 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Albert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almahairi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Babaei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bashlykov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhargava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhosale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bikel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Blecher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Ferrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Cucurull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Esiobu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fernandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fuller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Goswami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hartshorn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Inan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kardas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kerkez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Khabsa</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Kloumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Korenev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Koura</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Liskovich</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Mao</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Martinet</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Mihaylov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mishra</surname>
            , I. Molybog,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Poulton</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Reizenstein</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Rungta</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Saladi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Schelten</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Silva</surname>
            ,
            <given-names>E. M.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>X. E.</given-names>
          </string-name>
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Taylor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>J. X.</given-names>
          </string-name>
          <string-name>
            <surname>Kuan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Yan</surname>
            , I. Zarov,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kambadur</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Narang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Stojnic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Edunov</surname>
            , and
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Scialom</surname>
            . Llama 2:
            <given-names>Open</given-names>
          </string-name>
          <string-name>
            <surname>Foundation</surname>
          </string-name>
          and
          <string-name>
            <surname>Fine-Tuned Chat Models</surname>
          </string-name>
          .
          <year>2023</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2307.09288. arXiv:
          <volume>2307</volume>
          .09288 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,.
          <string-name>
            <surname>Kaiser</surname>
            ,
            <given-names>and I. Polosukhin.</given-names>
          </string-name>
          “
          <article-title>Attention Is All You Need”</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          . Vol.
          <volume>30</volume>
          . Curran Associates, Inc.,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          .
          <article-title>Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond</article-title>
          .
          <year>2023</year>
          . DOI:
          <volume>10</volume>
          .48550/arXiv.2304.13712. arXiv:
          <volume>2304</volume>
          .13712 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          1.
          <article-title>What was the population of Ireland in 1841?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>2. When did Irish Potato Famine start?</mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>3. When did Irish Potato Famine finish?</mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          4.
          <article-title>When did the Irish Parliament past the Act of Union?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>5. Where in Ireland are Protestants found?</mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <article-title>6. Where in Ireland are native Irish speakers found?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          7.
          <article-title>What was the population of Ireland in 1861?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>8. When did Catholic Emancipation happen?</mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          9. What happened in Ireland in
          <year>1848</year>
          ?
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          10.
          <article-title>When was the Land League created?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          1.
          <article-title>Can you make a definition of migration in ten sentences?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>2. What is Irish nationalism?</mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          3.
          <article-title>Can you summarize the three main arguments in Kerby Millers book Emigrants</article-title>
          and Exiles?
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          4.
          <article-title>Can you summarize the three main arguments in Donald Akensons book The Irish in Ontario?</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>5. What is the Irish diaspora?</mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>6. What is Irish unionism?</mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          7.
          <article-title>What are three differences between Irish nationalism</article-title>
          and Irish unionism?
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          8.
          <article-title>What are three similarities between Irish nationalism</article-title>
          and Irish unionism?
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          9.
          <article-title>Can you summarize what David Fitzpatricks book Oceans of Consolation is about?</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>