<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">ROGER: Extracting Narratives Using Large Language Models from Robert Gerstmann&apos;s Historical Photo Archive of the Sacambaya Expedition in 1928</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Mauricio</forename><surname>Matus</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Journalism</orgName>
								<orgName type="institution">Universidad Católica del Norte</orgName>
								<address>
									<settlement>Antofagasta</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Diego</forename><surname>Urrutia</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Computing &amp; Systems Engineering</orgName>
								<orgName type="institution">Universidad Católica del Norte</orgName>
								<address>
									<settlement>Antofagasta</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Claudio</forename><surname>Meneses</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Computing &amp; Systems Engineering</orgName>
								<orgName type="institution">Universidad Católica del Norte</orgName>
								<address>
									<settlement>Antofagasta</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Brian</forename><surname>Keith</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Computing &amp; Systems Engineering</orgName>
								<orgName type="institution">Universidad Católica del Norte</orgName>
								<address>
									<settlement>Antofagasta</settlement>
									<country key="CL">Chile</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">ROGER: Extracting Narratives Using Large Language Models from Robert Gerstmann&apos;s Historical Photo Archive of the Sacambaya Expedition in 1928</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">E37C6E53BBD87935DF014F9AD28DDA5E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Narrative Extraction</term>
					<term>Heritage Image Archives</term>
					<term>Sacambaya Expedition</term>
					<term>Large Language Models</term>
					<term>Image Labeling. 1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This article presents the ongoing work on developing a methodology for the systematic analysis and narrative of heritage image archives, focusing on the photo archive by Robert Gerstmann of the Sacambaya Expedition in 1928. This work combines state-of-the-art artificial intelligence techniques, such as advanced algorithms used in computer visión like convolutional neural networks, with Large Language Models (LLM) for generation purposes. The intent is to establish a practical and accessible framework in this area for institutions and individuals. The proposed method incorporates human-generated image labels with LLMs to produce narratives that aid researchers and users in their sense-making process while they explore a large archive of images. Through this iterative process, we aim to contribute not only to the understanding of this specific historical photo collection but also to the broader development of scalable solutions for the exploration and interpretation of heritage image archives. We seek to achieve a deeper understanding of the contents and meanings of the analyzed files, suggesting and highlighting new clustering of these materials and thematic/narrative connections that may not have been considered by a human observer.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The emergence of new technologies and the availability of a vast photographic archive have motivated a multidisciplinary project that explores the potential of these two elements in expanding the scope of multidisciplinary research. This ongoing research in the field of computational narrative extraction aims to develop a methodology for the analysis and semiautomatic construction of meaning and narratives from historical image archives using Large Language Models (LLMs). On the heritage side of this research, we aim to uncover and specify narratives inherent in large banks of photos for which there is limited information and dissemination <ref type="bibr" target="#b0">[1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">The Sacambaya Expedition photo archive</head><p>Robert Gerstmann (1896-1964) a German engineer and photographer, arrived in Chile in 1925. His work has been preserved in its original physical format since 1964 in Antofagasta, Chile. This material consists of 43,475 negatives and 15,054 positives in different formats, representing a period of photographic capture of approximately 40 years <ref type="bibr" target="#b1">[2]</ref>. The images detail landscapes from the beginning of the 20th century from the heights of Bolivia to Antarctica, the Pacific islands, and the Andes Mountain range <ref type="bibr" target="#b2">[3]</ref>. In this context, we focus on the Sacambaya Expedition archive, a part of the photographic work of Robert Gerstmann. In January 1928, Edgar Sanders, a Swiss engineer, established a company in London to search for an alleged Jesuit treasure hidden in an old monastery located in a ravine in the province of Inquisiví, Bolivia. The expedition comprised 21 individuals with diverse professional and military backgrounds, including 19 English citizens, 1 German, and 1 North American. The team scoured four different locations in the sector for five months, but their efforts proved fruitless, and they returned to Europe in November of that year <ref type="bibr" target="#b3">[4]</ref>. We note that most of the records related to this expedition have not been digitized before the present project. It is estimated that only approximately 15% of these images are digitized and accessible online <ref type="bibr" target="#b4">[5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">Proposed model</head><p>This research seeks to exploit the power of LLMs <ref type="bibr" target="#b5">[6]</ref>[7] and image processing techniques to extract narratives from a compendium of historical images. This endeavor aims to augment the corpus of historical knowledge by providing a narrative context to image archives that capture specific historical events. The overarching goal of this project is to establish a comprehensive framework/pipeline, named Robert Gerstmann Repository (ROGER), that enables efficient exploration, categorization, and semi-automated extraction of narratives implicit in previously unexamined heritage archives. The proposed methodology is designed to be iterative and incremental, ensuring clear documentation of progress through each phase.</p><p>The efficacy of the methodology is illustrated through a case study centered on a historical event, utilizing the framework to narrate its story systematically.</p><p>The ROGER Narrative Pipeline unfolds in a structured, multi-phased approach, incorporating human expertise and AI in a collaborative narrative pipeline. Initially, images are labeled through a combination of AI-driven algorithms and human judgment, resulting in a curated and contextually enriched dataset. Subsequent phases involve the use of AI to generate descriptive narratives and cluster these into thematic groups. The proposed pipeline integrates the use of prompt engineering loops <ref type="bibr" target="#b7">[8]</ref> -a process where human input is used to iteratively refine AI outputs -in the narrative extraction process, thus ensuring that the emerging narratives are not only accurate but also resonate with human interpretative frameworks. The final phase of the process is the drafting and construction of a coherent narrative with human feedback from the clustered image data. This interactive and iterative process between AI and human intelligence is instrumental in producing a polished and nuanced narrative output, ready for presentation and scholarly exploration.</p><p>This paper presents the results from the application of the ROGER Narrative Pipeline to the 1928 Sacambaya Expedition historical archive, uncovering the implicit narratives embedded within historical image archives. The proposed approach contributes to the interdisciplinary dialogue on narrative extraction, advancing our understanding of computational narrative construction in historical research. This represents a novel initiative to systematically decipher the stories enshrined in historical visual records. Finally, in the conclusion section, it is noted that this is the first work that has attempted to computationally unearth the stories contained in any of Robert Gerstmann's historical photo archives.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Technological framework</head><p>The computational analysis and interpretation of historical image archives is an emerging interdisciplinary field that integrates computer vision <ref type="bibr" target="#b8">[9]</ref>, natural language processing [10], information retrieval <ref type="bibr" target="#b10">[11]</ref>, and historical research methods <ref type="bibr" target="#b11">[12]</ref>. Several recent projects <ref type="bibr" target="#b12">[13]</ref> have demonstrated the potential of computational techniques to aid in making sense of largescale image archives and constructing historical narratives from them. From a technological point of view, the process described in the following Methodology section uses two key technologies, in order to implement a pipeline that takes a set of related patrimonial pictures as input and generates a coherent set of narratives as output. These key technologies are LabelBox and LLMs (e.g., ChatGPT).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Label Box</head><p>Labelbox is a machine learning annotation platform that simplifies the creation and management of annotated datasets, which are vital for AI development. It supports a variety of data types and provides tools for both manual and semi-automated annotation, aimed at increasing efficiency and accuracy. The platform encourages teamwork with its collaborative features and maintains high-quality annotations with robust quality control measures. Its user-friendly interface and integrative capabilities with machine learning workflows make it accessible for users of different skill levels, streamlining the annotation process from start to finish. Sensemaking of archives requires synthesizing across individual images to construct a higher-level understanding. Computational techniques for visual storytelling aim to build such narratives from image sequences <ref type="bibr" target="#b17">[18]</ref>. However, this typically relies on constrained domains with limited vocabularies <ref type="bibr" target="#b18">[19]</ref>. Our work leverages pre-trained LLMs capable of open-domain generation to construct narratives for historical image archives. In summary, our methodology builds upon advances in image recognition and LLMs while innovating in integrating these techniques for computational sensemaking over historical image archives. We believe this approach can provide both a macro-level narrative as well as a detailed understanding grounded in the image contents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">LLMs and ChatGPT</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>We present the ROGER Narrative Pipeline for the extraction of computational narratives from visual datasets. The general methodology is presented in Figure <ref type="figure" target="#fig_3">1</ref>. This process commences with a systematic labeling phase where a set of input images is semantically annotated using a combination of software tools and human oversight, producing a curated dataset. This curation involves the enrichment of the images with contextual metadata in the form of image labels. These labels enhance the depth and relevance of the information that will be used in narrative construction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1:</head><p>The ROGER Narrative Pipeline -A systematic framework illustrating the collaborative human-AI process for generating structured narratives from a dataset of images. The pipeline integrates human expertise with advanced AI algorithms for data curation, enrichment, and narrative construction, culminating in a coherent final narrative output Central to our pipeline is the integration of an LLM, which generates textual descriptions and titles from the enriched image data. These descriptions are the bedrock upon which the narrative structure is built. Subsequently, we use the LLM to perform clustering on the textual descriptions to organize the images into clusters or coherent thematic groups (Figure <ref type="figure" target="#fig_2">3</ref>), followed by the establishment of a timeline, ordering these clusters and images to create a draft narrative sequence. Integral to this process are the prompt engineering loops, where human operators iteratively refine the AI prompts based on the outputs to produce a final narrative (Figure <ref type="figure" target="#fig_4">4</ref>). This iterative process is pivotal, allowing for the human operator's critical and creative inputs to sculpt the narrative, ensuring structural and thematic integrity.</p><p>The final stages of the pipeline revolve around the transformation of the AI-generated timeline into a narrative draft. This draft undergoes a human-led finalization process, where narrative/theme experts refine the storyline, ensuring linguistic precision, narrative flow, and overall coherence. The result is a final narrative that provides a textual representation of the visual data in a narrative format. The final output contains the ordered list of images, their titles, their description, and the associated narrative. Furthermore, the output is accompanied by a detailed cluster list that provides an overview of the narrative elements and their organization, thereby offering transparency into the narrative structure and content. Through this methodical and collaborative approach, the pipeline achieves a high-fidelity narrative extraction from visual inputs, demonstrating the potential for robust human-AI collaboration in this domain.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>The main results associated with each stage of our pipeline, around a subset of 12 heritage images, are presented here.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Data curation and sampling</head><p>To demonstrate the capabilities of our methodology, we present the results on a representative subset of 12 pictures from the photo archive of the Sacambaya Expedition in 1928 <ref type="bibr">[20][21]</ref>. This archive consists of 545 negative originals in 10 x 15 format. From this collection, we discarded 45 photographs in the present analysis due to their defective nature and/or being over-or under-exposed, as they did not provide relevant information for labeling and subsequent categorization.</p><p>Prior to this study, this photographic archive was sustained by historical sources and the account of at least one of its members published in 1934 <ref type="bibr" target="#b22">[22]</ref>. The historical narrative suggested by these sources coincides with what is seen in the heritage images found in the archive. These materials consist of a distinctive first phase of photos made on board a ship, another category of materials showing means of transportation and human displacement tasks, to finally settle into a last general category of images that exhibit the efforts of an excavation and the logistics that this semi-industrial human operation entails <ref type="bibr" target="#b4">[5]</ref>. Therefore, these 500 photographs were pre-organized into the following thematic clusters:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Journey by ship: Subgroup identified as "London to Arica" (LTA) with 69 images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Journey by land: Subgroup identified as "Arica to Sacambaya" (ATS) with 187 images • Excavation sites: Subgroup identified as "Sacambaya" (SAC) with 244 images.</p><p>From each of these thematic groups we intentionally selected 4 images to exemplify the general progress of the expedition, the methodology employed, and suggest a temporal order in the narration of the journey.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Data enrichment through manual labeling</head><p>The enrichment of the original 500 images through the application of annotation tagging (both classification and object) by humans was facilitated by the use of the LabelBox <ref type="bibr" target="#b23">[23]</ref> platform. The use of automatic classification algorithms was considered; however, the decision was made to prioritize establishing a baseline developed by humans. The procedure is comprised of several distinct phases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.1">Dataset generation</head><p>A corpus of 500 images was curated, encompassing approximately 2.1GB of data. These images were standardized to a width of 1920 pixels and subsequently compressed using the cjpeg software <ref type="bibr" target="#b24">[24]</ref> to reduce their file size, culminating in a consolidated dataset of approximately 370MB. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.2">Ontology definition</head><p>The ontology for the annotation process was crafted following an initial analysis of the photographs. This ontology comprised two sets of annotations: General Classification and Specific Objects. The development of this ontology was a critical step in ensuring that the annotations would be comprehensive and consistently applied across the entire image set by the research assistants. The detailed categorization was designed to facilitate nuanced analysis and to support the study's objectives by providing rich, structured data.</p><p>The general classification elements consisted of 22 labels (such as trees, road, city, boat, excavation, beach, square, etc.) that were defined based on a prior visual analysis by the authors. The specific elements were grouped into categories: person, animal, object, transportation, and landscape. Each of these categories encompassed between 5 to 7 labels. This structured approach to classification allows for a detailed and organized analysis of the photographs. By having both general and specific categories, the researchers could ensure that the labeling process was thorough and nuanced, capturing both broad and fine-grained details within the images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.3">Annotator recruitment and annotation process</head><p>Ten undergraduate students from the fields of journalism and computer engineering were recruited to serve as research assistants. These individuals were selected based on the criterion of having completed at least 50% of their academic program and having prior participation in various research projects. Over the course of two weeks, these individuals were tasked with the systematic manual delineation of bounding boxes and the subsequent a) b)</p><p>assignment of labels corresponding to the identified objects within the images. Each photograph was labeled by at least two students and subsequently reviewed by the researchers. The team identified a total of 4,868 objects, of which 44.1% were classified as people, 22.17% as landscape details, 15.14% as various objects, with the remaining categories including animals and transportation. The average number of annotations produced by each student was 54.4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.4">Annotation extraction</head><p>The annotations were subsequently extracted from LabelBox in JSON format, which facilitated their processing and analysis through the Python programming language. In a detailed analysis of the dataset, 12 images were selected as representative samples from the collection of 500, and their annotations were obtained for further investigation. This structured approach to annotation not only enhances the reliability of the data but also ensures a level of granularity that is conducive to subsequent computational analysis.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Descriptions and clustering</head><p>Following the manual labeling process, we generated narrative descriptions of each picture in the dataset. These descriptions are generated using ChatGPT with GPT4. In particular, we prompt the model with minimal context about the Sacambaya Expedition, upload the image, and provide the human-generated labels described previously. Table <ref type="table" target="#tab_1">1</ref> illustrates the prompt design used to generate a narrative description and title for the picture shown in Figure <ref type="figure" target="#fig_1">2</ref> and the output generated by GPT4. We note that our annotators created the tags in Spanish as this was their native language, they were left untranslated in the prompt. We note that there was not a significant difference in translating them beforehand. The minimum context information supplied to the model was the name, year, and purpose of the expedition. The prompt included the names of some key places in the historical narrative (Bolivia and Sacambaya) and the name and nationality of the photographer (see Tables <ref type="table" target="#tab_1">1 to 3</ref>). Following the construction of all the narrative descriptions, we used another prompt to ask the LLM to cluster the images based on their content and reorganize them chronologically. The prompt used all the image descriptions generated beforehand to generate these clusters. The prompt is shown in Table <ref type="table" target="#tab_3">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Timeline and narrative draft</head><p>Following this clustering process, we asked the LLM to generate a timeline of the photos followed by a draft of the final narrative. We note that our work required extensive prompt design to ensure that the generated descriptions, clusters, and final narratives were coherent. We show the final version of the timeline and narrative extraction prompt alongside the corresponding narrative summary output in Table <ref type="table">3</ref>. The final timeline is shown in Figure <ref type="figure" target="#fig_4">4</ref> with the corresponding images. These two elements (timeline and narrative summary) represent the final output of the proposed pipeline. We note that our proposed prompt structure forces the model to hypothesize a plausible chronological order before generating the narrative draft. The data set of historical photos does not contain explicit temporal or spatial information not sharing with the model the thematic grouping mentioned in point 4.1. Therefore the LLM must be able to infer the order of the photos based on the general context of the expedition. While a human could also help with the timeline generation process and provide further information via prompting, the amount of photos in a full archive can make this generally unfeasible and too complex to scale. Thus, for scalability purposes, we rely on the capabilities of the LLMs to make sense of the underlying temporal and spatial information.</p><p>In general, sorting an unordered set of photos in the correct temporal order is a complex problem. In this context, GPT4 had issues when generating the correct timeline with only limited contextual information, which is why we had to include additional context in our prompt design. In reviewing the results, we found that the LLM model presented historical events in the correct chronological order, as long as it was provided with minimal contextual information. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Timeline extraction and narrative summarization prompt and final "Narrative Summary" output generated over the 12 photos subset of the historical archive. The timeline portion of the output is shown in Figure <ref type="figure" target="#fig_4">4</ref>. The prompt was followed by the list of all the previously generated descriptions of the photos.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt</head><p>You are an expert documentary filmmaker and you are exploring the historical archive of photos of the Sacambaya Expedition in Bolivia from 1928, taken by the German photographer Robert Gerstmann. The expedition sought to find the lost treasure of the jesuits, hidden somewhere near Sacambaya. You have the following list of descriptions of photos. Your goal is to generate two outputs: a timeline of these photos from the provided descriptions so that it forms a coherent narrative and a written summary of the narrative. The timeline should explicitly reference the IDs of the photos. The narrative summary should also reference the IDs of the photos as needed between brackets. Note that the provided descriptions might not be shown in the correct order and the IDs are not associated with any ordering. Please avoid the use of flowery language, your narrative should be written in an academic style. Do not explain your process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Output</head><p>The Sacambaya Expedition of 1928, documented by German photographer Robert Gerstmann, embarked on a challenging quest to uncover the Jesuit's lost treasure in Bolivia's Sacambaya Valley. It began with maritime scenes (IDs 65 and 60), capturing the tranquil ocean and coastal topography as the British expedition approached its destination. Aboard the ship, the crew is depicted in a moment of camaraderie and preparation (ID 30), followed by a brief sojourn at a riverine port, indicating the integration of transportation modes (ID 23). Upon reaching the Sacambaya Valley, the expedition faced the valley's rugged terrain (ID 40), fording rivers and navigating hilly landscapes (ID 85), and traversing dense thickets (ID 91). The photographs depict the challenging nature of their journey, highlighting both the physical and logistical hurdles encountered. As the expedition progressed, the team's burden of transporting heavy machinery through the dry valley is evident (ID 71), reflecting the blend of hope and hardship. Moments of rest and mechanical endeavors are captured (ID 46), as well as industrious activities amidst natural ruins (ID 37), underscoring the expedition's reliance on technology of the time. The concluding images (IDs 33 and 81) focus on the excavation and manual labor involved in the treasure hunt, set against a backdrop of natural and historical landscapes. This narrative, woven through Gerstmann's photographs, tells a story of determination, challenge, and the enduring human spirit in the pursuit of historical treasures.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>This ongoing research represents an advancement in the field of narrative extraction from heritage image archives. Through the iterative and incremental development of our methodology, a framework is established that not only enhances the understanding of historical events but also contributes to the broader discourse on computational narratives and artificial intelligence <ref type="bibr" target="#b25">[25]</ref>. This experimental model facilitates a utilitarian approach to solving issues prevalent in numerous audiovisual archives. The methodology proposed herein aims to achieve quality control measures concurrently with narration. While beyond this work's scope, future initiatives will prioritize expanding the search range and implementing an integral quality control system to regulate labeling accuracy, clustering thoroughness, and narrative coherence levels. Additionally, we acknowledge the potential of Large Language Models (LLMs) to autonomously identify a broad spectrum of objects. Consequently, subsequent projects may explore creating narrations without human support, relying solely on object identification and computer vision capabilities. However, if pursuing this avenue, methodological and future discussion elements must be incorporated to ensure a minimum level of confidence in the results, rendering them relevant for anthropological, historical, and heritage discourse. Limitations of this case study and the need for a more comprehensive evaluation are identified, as its primary objective was to illustrate the methodology rather than validate it with an extensive data spectrum.</p><p>The successes of the proposed method in constructing coherent historical narratives suggest a potential paradigm shift in how narrative extraction from visual historical records can be approached. Thus, this ongoing research represents a significant contribution to the challenge of uncovering narratives concealed within historical image archives. Furthermore, our aim is to observe significant changes, trends, and prevalent elements in large groups of visual information that may not be readily apparent through individual observations <ref type="bibr" target="#b26">[26]</ref>. This broader perspective facilitates the finding and construction of narratives that extend beyond individual images.</p><p>These experiments using easily available LLMs demonstrate the need to always maintain human control in the process, as shown by all the required prompt engineering. Future work will consist of applying the proposed pipeline to the entire collection of 500 images. We hope that our proposed methodology and technical pipeline streamline the work of expert catalogers, documentarians, and media creators, who can now have a minimal foundational basis to explore large, undisclosed photographic collections. Additionally, in the present case of the archive of the Sacambaya Expedition by Robert Gerstmann, we hope to have contributed to the historical and heritage enrichment of a part of this little explored collection.</p><p>In conclusion, we propose that the analysis of a specific photographic collection can be further enriched through the organization and utilization of information in narratives <ref type="bibr" target="#b12">[13]</ref>. Theories on sensemaking emphasize that sensemaking and narrative are two inherently interconnected concepts about how people understand the world around them <ref type="bibr" target="#b27">[27]</ref>. Given its replicability, we consider our proposed method to be a contribution to the discovery, enrichment, and dissemination of the worlds and narratives "hidden" inside photographic heritage archives.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>In: R. Campos, A. Jorge, A. Jatowt, S. Bhatia, M. Litvak (eds.): Proceedings of the Text2Story'24 Workshop, Glasgow (Scotland), 24-March-2024 mmatus@ucn.cl (M. Matus); durrutia@ucn.cl (D. Urrutia); cmeneses@ucn.cl (C. Meneses); brian.keith@ucn.cl (B. Keith) 0009-0003-9852-5285 (M. Matus); 0000-0002-0308-2406 (D. Urrutia); 0000-0003-1112-4925 (C. Meneses); 0000-0001-5734-8962 (B. Keith).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: a) Original photo in the context of the Sacambaya expedition; b) The same photo with objects identified by humans (rectangular green box, indicating its type and name from a predefined set of categories).</figDesc><graphic coords="5,112.32,252.24,366.00,156.72" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: ChatGPT proposal of coherent thematic groups clustered the images around common elements derived from manual labeling and descriptions of the images created with LLMs. We include the unsolicited titles created by GPT for the clusters.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Cluster 1 :</head><label>1</label><figDesc>"Maritime Prelude". Cluster 2: ""Expedition Life and Challenges". Cluster 3: "Industrial and Excavation Efforts".</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The timeline generated for the subset of 12 photographs orders and describes the European expedition in general terms, from its inception in Liverpool in March of that year to the depths of the Bolivian mountains between June and October 1928.</figDesc><graphic coords="8,121.20,432.24,351.36,255.36" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>An example of the prompt for AI-generated description.</figDesc><table><row><cell></cell><cell>Maquinaria -Personas -Árbol / Arbusto</cell></row><row><cell></cell><cell>Elementos Especificos: -Paisaje Tipo Árboles / Arbustos -Persona</cell></row><row><cell></cell><cell>Sexo Masculino Edad Adulto -Objeto Tipo Otro Nombre Cubeta -</cell></row><row><cell></cell><cell>Objeto Tipo Otro Nombre Barril -Transporte Tipo Tren /</cell></row><row><cell></cell><cell>Locomotora / Carro -Objeto Tipo Maquinaria -Paisaje Tipo</cell></row><row><cell></cell><cell>Árboles / Arbustos</cell></row><row><cell></cell><cell>Generate a short title for the photo and a description of the events</cell></row><row><cell></cell><cell>depicted by the photo. The description must be one paragraph</cell></row><row><cell></cell><cell>long. Please avoid flowery language, your description must be</cell></row><row><cell></cell><cell>written in an academic style.</cell></row><row><cell>Output</cell><cell>Title: "</cell></row></table><note>PromptYou are an expert documentary filmmaker and you are exploring the historical archive of photos of the Sacambaya Expedition in Bolivia from 1928, taken by the German photographer Robert Gerstmann. The expedition sought to find the lost treasure of the Jesuits, hidden somewhere near Sacambaya. You have the following picture that was tagged by humans in Spanish. Repeated tags mean that multiple objects of the same type appear in the image. Here are the tags: Elementos Generales: -Quebrada / Valle -Ruinas / Escombros -</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Excavation Effort in Sacambaya Valley, 1928"</head><label></label><figDesc></figDesc><table><row><cell>Description: This historical photograph captures a scene from the</cell></row><row><cell>1928 Sacambaya Expedition in Bolivia, overseen by German</cell></row><row><cell>photographer Robert Gerstmann. The image depicts a group of</cell></row><row><cell>adult male workers engaged in what appears to be an excavation</cell></row><row><cell>or construction activity within a quebrada or valley, surrounded</cell></row><row><cell>by the natural landscape of trees and shrubs. The men are</cell></row><row><cell>observed utilizing basic equipment and machinery of the era,</cell></row><row><cell>including buckets and barrels, indicative of manual labor</cell></row><row><cell>practices. A rudimentary transport system, possibly a narrow-</cell></row><row><cell>gauge railway, is also visible, suggesting the movement of</cell></row><row><cell>materials or debris. The terrain is littered with ruins and rubble,</cell></row><row><cell>perhaps remnants of previous structures or the outcome of the</cell></row></table><note>ongoing excavation. This photo serves as a visual record of the intense human and mechanical effort invested in the search for the Jesuits' lost treasure near Sacambaya, reflecting both the arduous conditions and the determination of the expedition members.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Clustering prompt and corresponding output. The prompt was followed by the list of all the previously generated descriptions of the photos. Only the relevant parts of the output are shown due to space constraints.PromptYou are an expert documentary filmmaker and you are exploring the historical archive of photos of the Sacambaya Expedition in Bolivia from 1928, taken by the German photographer Robert Gerstmann. The expedition sought to find the lost treasure of the jesuits, hidden somewhere near Sacambaya. You have the following list of descriptions of photos. Your goal is to identify clusters of the photos based on their content and their temporal order. Your second goal is to generate narrative descriptions of the identified clusters. Note that the provided descriptions might not be shown in the correct order and the IDs are not associated with any ordering. Please avoid the use of flowery language, the description of the clusters should be written in an academic style.</figDesc><table><row><cell>Output</cell><cell>Image IDs for each cluster.</cell></row><row><cell></cell><cell>Cluster 1 "Maritime Prelude": 23, 65, 60, 30.</cell></row><row><cell></cell><cell>Cluster 2 "Expedition Life and Challenges": 40, 91, 85, 71.</cell></row><row><cell></cell><cell>Cluster 3 "Industrial and Excavation Efforts": 46, 37, 81.</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The authors wish to acknowledge the contribution of the UCN Faculty of Humanities with its grant "Concurso de Incentivo Productividad Científica", 2023 initiative, which contributed financially to the project, and the UCN Library for allowing access to and work on Robert Gerstmann's photo archive. The authors also wish to thank the team of research assistants, made up of students from the School of Journalism and the Department of Computing and Systems Engineering, who carried out the task of manual classification and annotation of the more than 500 photos of the group under study.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Artificial Intelligence for Content and Context Metadata Retrieval in Photographs and Image Groups</title>
		<author>
			<persName><forename type="first">Peter</forename><forename type="middle">&amp;</forename><surname>Fornaro</surname></persName>
		</author>
		<author>
			<persName><surname>Chiquet</surname></persName>
		</author>
		<author>
			<persName><surname>Vera</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Archiving Conference</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="79" to="82" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Alvarado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roberto</forename><surname>Gerstmann</surname></persName>
		</author>
		<title level="m">fotografías, paisajes y territorios latinoamericanos</title>
				<meeting><address><addrLine>Pehuén, Santiago, Chile</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
	<note>1st. ed</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Roberto Gerstmann&apos;s last photography, Video</title>
		<author>
			<persName><forename type="first">M</forename><surname>Matus</surname></persName>
		</author>
		<ptr target="https://youtu.be/9nFvhoZd5Os" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">The Story of the Jesuit Gold Mines in Bolivia and of the Treasure Hidden by the Sacambaya River</title>
		<author>
			<persName><forename type="first">E</forename><surname>Sanders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename></persName>
		</author>
		<imprint>
			<date type="published" when="1928">1928</date>
		</imprint>
		<respStmt>
			<orgName>Rauner Special Collections Library -Dartmouth College</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Pavez</surname></persName>
		</author>
		<title level="m">Imágenes de la revolución industrial: Robert Gerstmann en las Minas de Bolivia</title>
				<meeting><address><addrLine>Plural, La Paz, Bolivia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1925">1925-1936. 2017</date>
		</imprint>
	</monogr>
	<note>1st ed</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Semantic clustering based deduction learning for image recognition and classification</title>
		<author>
			<persName><forename type="first">Wenchi</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xuemin</forename><surname>Tu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bo</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guanghui</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">124</biblScope>
			<biblScope unit="page">108440</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Large language models: Their success and impact</title>
		<author>
			<persName><forename type="first">Spyros</forename><surname>Makridakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fotios</forename><surname>Petropoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yanfei</forename><surname>Kang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Forecasting</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="536" to="549" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Prompt Engineering or Fine Tuning: An Empirical Assessment of Large Language Models in Automated Software Engineering Tasks</title>
		<author>
			<persName><forename type="first">Jiho</forename><surname>Shin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Clark</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tahmineh</forename><surname>Mohati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maleknaz</forename><surname>Nayeb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Song</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hadi</forename><surname>Hemmati</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024-10">2024. October 2024</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">22</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">What to do with 2.000.000 historical press photos? The challenges and opportunities of applying a scene detection algorithm to a digitised press photo collection</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wevers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Vriend</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>De Bruin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">TMG Journal for Media History</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Integrating wiki systems, natural language processing, and semantic technologies for cultural heritage data management</title>
		<author>
			<persName><forename type="first">R</forename><surname>Witte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kappler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Krestel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">C</forename><surname>Lockemann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Language Technology for Cultural Heritage: Selected Papers from the LaTeCH Workshop Series</title>
				<meeting><address><addrLine>Berlin Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="213" to="230" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Annotating historical archives of images</title>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Keogh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Shelton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th ACM/IEEE-CS joint conference on Digital libraries</title>
				<meeting>the 8th ACM/IEEE-CS joint conference on Digital libraries</meeting>
		<imprint>
			<date type="published" when="2008-06">2008. June</date>
			<biblScope unit="page" from="341" to="350" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">S</forename><surname>Jo</surname></persName>
		</author>
		<title level="m">Foreign Relations of the United States Series, 1860-1980: A Study in New Archival History</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
		<respStmt>
			<orgName>Stanford University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Storytelling with Image Data: A Systematic Review and Comparative Analysis of Methods and Tools</title>
		<author>
			<persName><forename type="first">F</forename><surname>Lotfi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Beheshti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Farhood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pooshideh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jamzad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Beigy</surname></persName>
		</author>
		<idno type="DOI">10.3390/a16030135</idno>
		<ptr target="https://doi.org/10.3390/a16030135" />
	</analytic>
	<monogr>
		<title level="j">Algorithms</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page">135</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Language Models Are Few-shot learners</title>
		<author>
			<persName><forename type="first">T</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Kaplan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances In Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="1877" to="1901" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">PaLM: Scaling language modeling with pathways</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chowdhery</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">W</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Sutton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gehrmann</surname></persName>
		</author>
		<idno>arXivpreprint2204.02311</idno>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">LLaMA: Open and efficient foundation language models</title>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Rozi`ere</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hambro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Azhar</surname></persName>
		</author>
		<idno>arXivpreprint2302.13971</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title/>
		<idno type="DOI">10.48550/arXiv.2303.08774</idno>
		<idno>report. arXivpreprint2303.08774</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2303.08774" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
		<respStmt>
			<orgName>OpenAI</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">GPT-4 technical</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Glac net: Glocal attention cascading networks for multi-image cued story generation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">O</forename><surname>Heo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Son</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">W</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">T</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1805.10973</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">H</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ferraro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mostafazadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Misra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Visual storytelling</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mitchell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies</title>
				<meeting>the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies</meeting>
		<imprint>
			<date type="published" when="2016-06">2016. June</date>
			<biblScope unit="page" from="1233" to="1239" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Highresolution aerial image labeling with convolutional neural networks</title>
		<author>
			<persName><forename type="first">Emmanuel</forename><surname>Maggiori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuliya</forename><surname>Tarabalka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guillaume</forename><surname>Charpiat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pierre</forename><surname>Alliez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Geoscience and Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="7092" to="7103" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Tales of Glitter or Dust</title>
		<author>
			<persName><forename type="first">D</forename><surname>Buck</surname></persName>
		</author>
		<ptr target="https://www.thefreelibrary.com/TalesofGlitterorDust.-a073064246" />
		<imprint>
			<date type="published" when="2000-12">2000. December 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">The Treasure Trail</title>
		<author>
			<persName><forename type="first">S</forename><surname>Jolly</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1934">1934</date>
			<publisher>John Long Limited</publisher>
			<pubPlace>London</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Labelbox</title>
		<author>
			<persName><surname>Labelbox</surname></persName>
		</author>
		<ptr target="https://labelbox.com" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">The JPEG Still Picture Compression Standard</title>
		<author>
			<persName><forename type="first">Gregory</forename><forename type="middle">K</forename><surname>Wallace</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="30" to="44" />
			<date type="published" when="1991-04">April 1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">A survey on event-based news narrative extraction</title>
		<author>
			<persName><forename type="first">Keith</forename><surname>Norambuena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Brian</forename><surname>Felipe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tanushree</forename><surname>Mitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chris</forename><surname>North</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="issue">14s</biblScope>
			<biblScope unit="page" from="1" to="39" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">The civilizing process in London&apos;s Old Bailey</title>
		<author>
			<persName><forename type="first">Sara</forename><forename type="middle">&amp;</forename><surname>Klingenstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tim</forename><forename type="middle">&amp;</forename><surname>Hitchcock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Simon</forename><surname>Dedeo</surname></persName>
		</author>
		<idno>pnas.1405984111</idno>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the National Academy of Sciences of the United States of America</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<biblScope unit="page">10</biblScope>
			<date type="published" when="1073">2014. 1073</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A System for Image Understanding using Sensemaking and Narrative</title>
		<author>
			<persName><forename type="first">Zev</forename><surname>Battad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mei</forename><surname>Si</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Ninth Advances in Cognitive Systems (ACS) Conference</title>
				<imprint>
			<date type="published" when="2021">2022. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
