<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automatic knowledge-graph creation from historical documents: The Chilean dictatorship as a case study</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Camila</forename><surname>Díaz</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jocelyn</forename><surname>Dunstan</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Pontificia Universidad Católica de Chile</orgName>
								<address>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lorena</forename><surname>Etcheverry</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Universidad de la República</orgName>
								<address>
									<country key="UY">Uruguay</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Antonia</forename><surname>Fonck</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Pontificia Universidad Católica de Chile</orgName>
								<address>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alejandro</forename><surname>Grez</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Domingo</forename><surname>Mery</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Pontificia Universidad Católica de Chile</orgName>
								<address>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Juan</forename><surname>Reutter</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Pontificia Universidad Católica de Chile</orgName>
								<address>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hugo</forename><surname>Rojas</surname></persName>
							<affiliation key="aff0">
								<address>
									<settlement>IMFD</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="institution">Universidad Alberto Hurtado</orgName>
								<address>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
							<affiliation key="aff4">
								<address>
									<settlement>VioDemos</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automatic knowledge-graph creation from historical documents: The Chilean dictatorship as a case study</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">7582DB4F29C07EBBD4E5763AD11E9FD9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We present our results regarding the construction of a knowledge graph from historical documents related to the Chilean dictatorship period (1973)(1974)(1975)(1976)(1977)(1978)(1979)(1980)(1981)(1982)(1983)(1984)(1985)(1986)(1987)(1988)(1989)(1990). Our approach uses LLMs to automatically recognize entities and relations between them and resolve conflicts between these values. To prevent hallucination, the interaction with the LLM is grounded in a simple ontology with four types of entities and seven types of relations.</p><p>To evaluate our architecture, we use a gold standard graph constructed using a small subset of the documents, and compare this to the graph obtained from our approach when processing the same set of documents. Results show that the automatic construction manages to recognize a good portion of all the entities in the gold standard and that those not recognized are explained mainly by the level of granularity in which the information is structured in the graph and not because the automatic approach misses an important entity in the graph. Looking forward, we expect this report to encourage work on other similar projects focused on enhancing research in humanities and social science. However, we remark that better evaluation metrics are needed to accurately fine-tune these types of architectures.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Knowledge graphs have been identified as a promising tool for analyzing historical documents <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>. Indeed, given a collection of documents, one can build a knowledge graph by identifying all relevant entities and relations between them. The construction and usage of such knowledge graphs allow shifting focus from a document-centric approach, in which analysts must find their information within a collection of documents, to an entity-centric approach, in which users can immediately find relevant entities in knowledge graphs, together with other helpful information such as the neighborhood of these entities, the relation or paths between them, and other more complex patterns.</p><p>While promising, the construction of these knowledge graphs is a challenging, expensive endeavor. To identify entities, one must read all relevant documents, and this list must be constantly curated to avoid duplication. The same applies to the discovery of the relations between entities. One way of partially preventing the cost of constructing knowledge graphs is to look for automatic or semi-automatic techniques in which one leverages recent advances in Natural Language Processing (NLP) to take care of the construction of the knowledge graph or to deliver a preliminary result that is then later manually curated at a much lower cost.</p><p>In this paper, we present results regarding the automatic construction of a knowledge graph containing information about the human rights violations committed by the Chilean dictatorship of Augusto Pinochet between 1973 and 1990. The transition to democracy and transitional justice began in Chile after the peaceful return to democracy in 1990. In these 34 years, there has been notable progress in the five different elements of transitional justice-the search for truth, justice, reparation, memory, and non-repetition-although important challenges remain pending <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref>.</p><p>For the work presented here, we use documents retrieved from the digital archive Memoria Viva <ref type="foot" target="#foot_0">1</ref> . This repository is part of the NGO Human Rights International Project, initially established by Chilean refugees and human rights activists in London. Their goal is to collect and make available to a broad public the documentation that records the crimes committed by the dictatorial state apparatus, relying on primary and secondary sources such as Official Reports, Judicial and Police Archives, Archives of Human Rights Organizations, Testimonials, Journalistic Documentation/Books, and International Archives Our ultimate goal is the creation of innovative computational methods specifically designed for analyzing historical documents, which will enable the development of tools to integrate currently fragmented information while adhering to necessary quality and standards. This endeavor is also intended to impact the process of building historical knowledge, providing a space for the interrelation between fragments that enable more integrated and comprehensive analyses, supporting the work of different disciplines and organizations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related work</head><p>Information extraction tasks are usually divided into open and closed tasks. Open information extraction (OIE) is designed to derive relation triplets from unstructured text by directly using entities and relations from the sentences without adhering to a fixed schema. In contrast, closed information extraction (CIE) focuses on extracting factual data from text that conforms to a predetermined set of relations or entities, as detailed by <ref type="bibr" target="#b5">[6]</ref>. Our approach can be classified under the CIE paradigm. Still, we use a simple domain ontology to guide the process of extracting entities and relationships between entities instead of a predefined set of concrete entities or relations.</p><p>Regarding the machine learning models to extract entities and relationships, several papers in recent years have explored comprehensive methods that use a single machine learning model for joint-named entity recognition and relationship extraction (NERRE) <ref type="bibr" target="#b6">[7]</ref>. These methods typically take a sequence-by-sequence approach in which a model is trained to obtain tuples of two or more named entities and the relationship label belonging to the predefined set of possible relationships between them. In <ref type="bibr" target="#b7">[8]</ref>, the authors present an approach using LLMs for NERRE in the materials science domain, extracting hierarchical entities and relationships between them. Our work with historical documents presents new challenges to this task: the documents we process contain various information, ranging from judicial processes to family relationships. This means the extraction process must be general enough to accommodate general information instead of focusing on a particular domain.</p><p>Although open and closed information extraction vary, both attempt to convert unstructured text into structured knowledge, usually represented in triples that help outline relationships but offer limited entity-level knowledge. It is often assumed that two triples refer to the same entity if their topics coincide. However, this assumption may not necessarily hold in practice. Furthermore, evaluating these tasks is based on precision, recall, and F1 at the triplet level, which can lead to erroneous conclusions about entity understanding. Some recent works propose new metrics to evaluate the quality of the results obtained beyond these classical metrics. Among them, the one defining the Approximate Entity Set OverlaP (AESOP) metric stands out <ref type="bibr" target="#b8">[9]</ref>. In this paper, we take inspiration from this work to find metrics suitable for comparing graphs produced by two possibly different black-box architectures. We only allow comparison at the level of the final answer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Our approach</head><p>Our approach is summarized in Fig. <ref type="figure" target="#fig_0">1</ref>. The idea is to repeatedly prompt an LLM for entities and relations and resolve these relations. This results in a graph, which is later post-processed to remove additional redundancy and fix possible mistakes in the previous steps. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">A simple (fixed) ontology</head><p>Our architecture considers a simple ontology in which we fix certain types of entities and relationships. We then carefully design prompts to extract these entities and relationships, using the concepts of the ontology. We do this to try to reduce the hallucinations of the LLM. However, we keep the ontology general enough not to lose valuable information. The proposed entities and relations are presented in Table <ref type="table">3</ref>.1. Furthermore, each relation type also restricts the type of its origin and target entities. In the rest of the section, we omit clear types from the context. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Entities</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Extracting Entities and Relations</head><p>Entities and relations are extracted by prompting LLMs. We use OpenAI's API 2 . The approach used is zero-shot prompting, as no examples of the expected output are included. However, the expected behavior and expected structure are detailed in a JSON document. We extract each type of entity or relation independently, with a different, manually created, prompt that broadly specifies the context of the documents, the data expected from the response, and the expected structure.</p><p>For example, the prompt to recognize entities of type Individual is as follows:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt to extract Individuals</head><p>Prompt: Your goal is to identify all the individuals mentioned in the document and provide the information about that person as a structured object. You will receive a document related to the Chilean dictatorship of 1973.</p><p>Generate a new JSON object containing the identified individuals:</p><p>{ "individual": [ // a list of all the individuals { "firstName": string, // first name of the individual "lastName": string, // last name of the individual "role": string, // individual's job, role, profession or activity. If not specified, make it "unspecified" "summary": string, // 1-sentence summary of the individual "origin_reference": string[] // If the document has parts identified with the field "ORIGIN_REFERENCE", here comes that value; if more than one corresponds, add them all as a list of strings } ] } Make sure that each individual found satisfies this sentence: "individual person with first and last name". Before adding the individual to the result, imagine a detailed explanation for why you deduce that the individual satisfies the sentence. If the explanation is not 100% convincing, ignore that result. Use only lowercase letters without accents. Use the English language for the summary.</p><p>As per relations, recall that we restrict each relation to the types of entities where they participate. Then, for every relation type 𝑅 between entities of type 𝐸 1 and 𝐸 2 , we form a prompt that concatenates all entities of type 𝐸 1 and 𝐸 2 discovered in the previous process and asks the LLM to discover relations between these entities only. For example, here is the prompt for the relation of type IndividualIsRelatedToOrganization (to which we must add the list of entities of type Individual and Organization that were previously discovered in the document):</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt to extract Individuals</head><p>Prompt: You will be given a JSON with a list of individuals (denoted between tags '==LIST 1 START==' and '==LIST 1 END=='), a JSON with a list of organizations (denoted between tags '==LIST 2 START==' and '==LIST 2 END=='), and then a document (denoted between tags '==DOCUMENT START==' and '==DOCUMENT END=='). Based on the context, you can identify from the document, find the individuals that are related to an organization somehow. The result must be a JSON object with the following structure: { "IndividualIsRelatedToOrganization": [ { "nature": string, // The nature of the relation of the individual with the organization, it must be "affected by", "member", "chief", or "other" "individualId": number, // ID of the individual that is a member of the organization "organizationId": number, // ID of the organization "summary": string, // 1-sentence summary of the relation "origin_reference": string[] // If the document has parts identified with the field "ORIGIN_REFERENCE", here comes that value; if more than one corresponds, add them all as a list of strings } ] } Each relationship should include a summary describing the nature of the relationship between the entities. Ensure meaningful relations are included, avoiding duplicates.</p><p>Use only lowercase letters without accents. Use the English language for the summary.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">The full text-to-graph pipeline</head><p>In the following, we explain the entire pipeline that we use to go from text to a knowledge graph, which we divide into three main steps. First, we need to split the text documents into blocks of smaller length to fit into the prompt we send to OpenAI's API. Then, we create a different prompt to extract each different entity type and relation, and use this prompt to extract entities and relation from each of these blocks. Lastly, we process this entity/relation graph to remove redundant or incorrect information based on the graph's structure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.1.">Document Splitting</head><p>Since we use OpenAI's GPT API, which has a size limit for its prompts input and output, we need to divide the text document into fragments of a smaller size. Moreover, we must choose a size that produces a good enough response. Generally, with a too small size, the LLM may not know what to look for and may produce forced or incorrect results. Conversely, with a size that is too big, the response may include only some occurrences. We tested lengths of 1000, 2000, 5000, 10000, and 15000 characters, making each cut either at a line break or a sentence end. We also added an overlay at the beginning and end of each fragment of 0.1 of its size to provide the LLM with more context. After trying different values, we found that the best results were obtained with fragments of 5000 characters with overlays of 500 characters.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.2.">Prompting to extract Entities and Relations</head><p>After splitting the documents into fragments, the next step is to extract each fragment's entities and relations using OpenAI's GPT API again. To extract the entities of a fragment for each entity type, we prompt the API with the corresponding prompt as the system role and the fragment as the user role and receive a JSON object with the entities of that type found in that fragment. Similarly, to extract the relations of a fragment, we prompt the API with the corresponding prompt for each relation type and add all entities found in that fragment that are of the types relevant to the relation, i.e., the types of the subject and object entities. We consider only one subject type and one object type for each relation. With this prompt, we get a JSON object with the relations of that type found in that fragment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.3.">Performing Entity Resolution</head><p>Next comes the task of finding duplicated entities, for which we merge both of their JSON files. More precisely, we apply a series of rules that select possible pair of duplicated entities. For each of these pairs, we ask an LLM to merge the information contained in both their JSON files.</p><p>To name some of the rules: for individuals, we consider duplicated any two entities A and B for which both the name and surname of A are equal to or contained in the name and surname of B, respectively. As another example, for relations between entities of type Location, we remove any containment relation between locations that do not follow the order country &gt; city &gt; street &gt; building <ref type="foot" target="#foot_1">3</ref> .</p><p>After this, we obtain our raw graph.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.4.">Graph Post-Processing</head><p>The raw graph is also post-processed to remove incorrect and redundant information by applying rules to its structure. First, we focus on edge removal. Generally, it would not make sense to have loops in this kind of graph, so we assume it is due to errors at the time of relation extraction. To fix it, we remove any existing relation participating in a loop (we use a loop limit of size 5 to ease the computation). The next step is to remove redundant edges that can be easily extracted from our ontology. For this, we define four rules that declare redundant relations, which can then be removed.</p><formula xml:id="formula_0">1. If 𝑂1 IsPartOf 𝑂2 and 𝑂2 IsPartOf 𝑂3, then 𝑂1 IsPartOf 𝑂3 is redundant; 2. I f 𝐸 OccursAt 𝐿1 and 𝐿1 IsContainedIn 𝐿2, then 𝐸 OccursAt 𝐿2 is redundant; 3. I f 𝐼 WasPresentAt 𝐿1 and 𝐿1 IsContainedIn 𝐿2, then 𝐼 WasPresentAt 𝐿2 on the same date is redundant; 4. I f 𝑂1 IsRelatedTo 𝐸 and 𝑂1 IsPartOf 𝑂2, then 𝑂2 IsRelatedTo 𝐸 is redundant.</formula><p>For readability, we have shortened up the relation names. The fouth rule applies to the properties IndividualIsRelatedToOrganization, IndividualIsRelatedToEvent or OrganizationIsRe-latedToEvent.</p><p>After removing incorrect and redundant relations, we focus on merging redundant nodes <ref type="foot" target="#foot_2">4</ref>Our initial experiments showed that the LLM sometimes confuses organizations and locations, resulting in duplicate entities that must be resolved. Therefore, we address this issue by identifying locations and organizations with the same name and resolving them into one entity: a location, an organization, or both. The latter case makes sense in specific instances, such as when an organization is located in a particular building, and the name could refer to both the organization and the building. We decide this based on which node has more information. The API annotates organization and location entities with a possibly unspecified nature parameter. If one node has this parameter defined and the other does not, then that node's type is prioritized; if both nodes have the value or neither does, then the node keeps both types.</p><p>The last step of our process is to merge duplicate event nodes. We aim to reduce the number of event nodes by applying a limit to its granularity in two ways. Note that every event is annotated with a property date that contains the date it takes place. First, we restrict that there can only be one event per location per date, so all events with the same values are merged into one. Second, an individual can be related to only one specific event on a particular date, so all events that share these values must be merged. Since applying any of the two rules might raise new ones for the other, we use them iteratively until no changes are made to the graph.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Evaluation</head><p>In this section we evaluate the quality of the proposed graph. Regarding the scalability of the approach, we note that our algorithms have been put to use in a real life scenario comprising approximately 40.000 pages of text, generating a graph with more than 50.000 entities and 100.000 relations in about two hours, spending approximately 30 usd in requests to the chatGPT API, according to prices published by the time of writing.</p><p>To validate the quality of the graph, we propose to examine documents from public records collected from a known collection (in our case, memoriaviva.com). We asked a domain expert to construct a gold-standard graph from a subset of the documents in this collection, totalling around 7500 words, and we compared how we fare against this standard graph.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">A gold standard sub-graph</head><p>Our gold standard graph relates to the facts presented in memoriaviva.com concerning the location of Lonquen, Chile, wherein a series of corpses were found during the dictatorship. The constructed graph contains 121 entities: 51 Individuals, 38 Events, 16 Locations, and 16 Organizations. When writing this paper, the domain expert was still resolving conflicts regarding some of the discovered relations, so we prefer not to report metrics on relations at this stage, leaving them for the camera-ready version.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head><p>To analyze how we fared on the task of recognizing entities in the graph, we count the number of Entities of each type that were recognized adequately by our automatic approach (Present in both graphs), as well as the number of entities that were extracted by the automatic approach but where not included in the gold standard graph by the domain expert (Extra nodes not in gold standard), and the number of entities we failed to identify (Missing nodes from gold standard). Because of the sheer number of nodes, we again resorted to LLM for this evaluation and asked LLMs to decide which nodes matched and which did not. Results are presented in table 4.2.</p><p>Interestingly, the precision of our algorithm varies tremendously depending on the entity's nature. The algorithm extracts individuals with high precision, and the only two missed nodes from the gold standard correspond to spouses of people who were reported to be detained and mentioned only once in all documents.</p><p>For the case of organizations, the comparison reveals a slight problem with our algorithm, as the automatic graph creates a few extra organizations. This happens when the same organization is mentioned in the document several times with different roles (i.e., a specific division of the police is mentioned to detain someone, and then this particular division is said to be led by a specific officer). In this case, our automatic construction would generate different entities for the various natures of the organizations (so one node corresponds to the police described as an organization that detains people, and another to the police described as an organization led by this officer).</p><p>Locations can be explained by a completely different phenomenon: our automatic graphs create a more detailed location graph than the one made by experts. For example, if an event is said to happen on the street 𝑆 of city 𝐶, experts only mention the city 𝐶, but our graph mentions both cities and streets. This suggests that our automatic graph could further refine the gold standard.</p><p>Finally, we have a similar problem with locations in the case of events. Still, in the opposite direction, experts provide much more detailed information about specific events, whereas our automatic extraction tends to merge two or three events into a single one. For example, experts would detail all the events leading to the detention of victims, including the order in which individuals were detained, how they were transported, and in which vehicle; the automatic graph would instead merge all these events into a single detention event concerning all individuals, or sometimes two events, one for the detention and another for the transportation of the victims.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions and lessons learned</head><p>The work presented here was a fully automatic analysis of a historical archive. We use large language models to recognize entities, perform entity resolution, establish relations, and resolve conflicts between identified relations. Importantly, we ensured that all the steps in our process preserve the linkage of identified entities and relations. The final knowledge graph preserves pointers to the original documents from which a given entity or relation was extracted (note that a single entity or relation may point to several documents, as entities and relations we identify are subject to a resolution step).</p><p>One crucial issue in this work is to measure the quality of the final knowledge graph. As several steps are involved, a slight change in the prompt used for any one step can create a massive difference in the resulting knowledge graph. Our approach was to measure the final product against a gold-standard graph. To test our algorithms, we first asked experts to manually construct a graph from a small sample of Living Memory documents (121 entities). Then, we compared the graph obtained by our approach, considering the same subset of texts, with the expert graph.</p><p>The resulting algorithms show promising results. Measured against the gold standard, the extraction of Individuals was extremely precise. On the other hand, the extraction of Organization, Events, and Location was less successful, but on closer inspection, this can be partially explained by differences in the granularity of the location and events, and one can verify that the automatic approach captures the big picture of events just as well as the gold standard. It remains to be seen whether this difference in granularity can be fixed by using better prompts or if we need new components in our architecture.</p><p>In terms of future work, we plan to systematically evaluate the entities detected by chatGPT. To do so, we want to create an annotated corpus with the entities of interest <ref type="bibr" target="#b9">[10]</ref>. Even if the annotation is time-consuming, it can evaluate generative AI and other possible approaches to detect critical information, such as Named Entity Recognition algorithms <ref type="bibr" target="#b10">[11]</ref>.</p><p>Building a knowledge graph from collections of historical archives related to the Chilean dictatorship is a way to add robust data science analyses to this critical event in Chilean history that still has so many open questions. Unlike manual analysis of fragmented archives, a unified approach can provide a comprehensive view, filling information gaps, cross-validating data, and enhancing contextual understanding. A holistic method can uncover more profound insights into the complexities of the dictatorship, enabling researchers to answer more nuanced questions and revealing connections and patterns that fragmented, manual analyses might miss. Of course, the quality of our analysis depends on the data quality. Thus, future research should resort to primary sources to complement and validate the compilation made by Memoria Viva.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Summary of the proposed approach.</figDesc><graphic coords="3,89.29,415.48,416.67,84.67" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Comparison of the entities in the automatically generated and gold standard graphs.</figDesc><table><row><cell>Entity Type</cell><cell></cell><cell></cell></row><row><cell></cell><cell>Present in both graphs</cell><cell>47</cell></row><row><cell>Individual</cell><cell>Extra nodes not in gold standard</cell><cell>0</cell></row><row><cell></cell><cell>Missing nodes from gold standard</cell><cell>2</cell></row><row><cell></cell><cell>Present in both graphs</cell><cell>14</cell></row><row><cell>Organization</cell><cell>Extra nodes not in gold standard</cell><cell>21</cell></row><row><cell></cell><cell>Missing nodes from gold standard</cell><cell>2</cell></row><row><cell></cell><cell>Present in both graphs</cell><cell>14</cell></row><row><cell>Location</cell><cell>Extra nodes not in gold standard</cell><cell>6</cell></row><row><cell></cell><cell>Missing nodes from gold standard</cell><cell>2</cell></row><row><cell></cell><cell>Present in both graphs</cell><cell>18</cell></row><row><cell>Event</cell><cell>Extra nodes not in gold standard</cell><cell>3</cell></row><row><cell></cell><cell>Missing nodes from gold standard</cell><cell>20</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The digital archive is available at https://memoriaviva.com/nuevaweb/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">We can do this because we specifically ask the LLM to specify the nature of the location at the time of entity recognition</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">The LLM annotates every entity with a summary property, and every time a pair of entities are merged, the resulting one maintains both summaries in a list. The same is done for relations.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Knowledge graphs</title>
		<author>
			<persName><forename type="first">C</forename><surname>Gutierrez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Sequeda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">64</biblScope>
			<biblScope unit="page" from="96" to="104" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Creating a knowledge graph for ireland&apos;s lost history: Knowledge engineering and curation in the beyond 2022 project</title>
		<author>
			<persName><forename type="first">C</forename><surname>Debruyne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Munnelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kilgallon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>O'sullivan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Crooks</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Journal on Computing and Cultural Heritage (JOCCH)</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="1" to="25" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Induction of a large-scale knowledge graph from the regesta imperii</title>
		<author>
			<persName><forename type="first">J</forename><surname>Opitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Born</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Nastase</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences</title>
				<meeting>the Second Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences</meeting>
		<imprint>
			<publisher>Humanities and Literature</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="159" to="168" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Rojas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shaftoe</surname></persName>
		</author>
		<title level="m">Human rights and transitional justice in Chile</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">50 years after the 1973 coup in chile: Analysis of the processes of transition to democracy and transitional justice</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">R</forename><surname>Corral</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Seattle J. Soc. Just</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page">587</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Genie: Generative information extraction</title>
		<author>
			<persName><forename type="first">M</forename><surname>Josifoski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">D</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Peyrard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>West</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.17617</idno>
		<title level="m">Large language models for generative information extraction: A survey</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Structured information extraction from scientific text with large language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Dagdelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dunn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Walker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Rosen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ceder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">A</forename><surname>Persson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jain</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41467-024-45563-x</idno>
		<ptr target="https://www.nature.com/articles/s41467-024-45563-x.doi:10.1038/s41467-024-45563-x" />
	</analytic>
	<monogr>
		<title level="j">Nature Communications</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="14" />
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Mikaelyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Meulemans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hensman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mitra</surname></persName>
		</author>
		<title level="m">Learning to extract structured entities using language models</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Collaborative annotation for reliable natural language processing: Technical and sociological aspects</title>
		<author>
			<persName><forename type="first">K</forename><surname>Fort</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>John Wiley &amp; Sons</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Simple yet powerful: An overlooked architecture for nested named entity recognition</title>
		<author>
			<persName><forename type="first">M</forename><surname>Rojas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bravo-Marquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dunstan</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.coling-1.184" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th International Conference on Computational Linguistics, International Committee on Computational Linguistics</title>
				<meeting>the 29th International Conference on Computational Linguistics, International Committee on Computational Linguistics<address><addrLine>Gyeongju, Republic of Korea</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="2108" to="2117" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
