<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Linking Dutch World War II Cultural Heritage Collections with Events Extracted by Machines and Crowds</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jessie</forename><surname>Both</surname></persName>
							<email>jessieboth@gmail.com</email>
						</author>
						<author>
							<persName><forename type="first">Ramses</forename><surname>Ijff</surname></persName>
							<email>ramses.ijff@gmail.com</email>
						</author>
						<author>
							<persName><forename type="first">Oana</forename><surname>Inel</surname></persName>
							<email>oana.inel@vu.nl</email>
						</author>
						<author>
							<persName><forename type="first">Victor</forename><surname>De Boer</surname></persName>
							<email>v.de.boer@vu.nl</email>
						</author>
						<author>
							<persName><forename type="first">Lora</forename><surname>Aroyo</surname></persName>
							<email>lora.aroyo@vu.nl</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">University of Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Didi de</orgName>
								<orgName type="institution">Hooge University of Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff5">
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Linking Dutch World War II Cultural Heritage Collections with Events Extracted by Machines and Crowds</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">655518A1B485B3265AB739A03AF3B45E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:02+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract/>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>The digital revolution is changing the way people access and use knowledge and information. By digitizing collections, cultural heritage institutions aim to fulfill their new role as information providers. However, the majority of the digitized collections lack sufficient metadata and provide no context. This makes the collections hard to access and to search through. Prior research <ref type="bibr" target="#b2">[3]</ref> shows that enriching metadata of digitized collection items with events helps with providing context, structure collections, and describe relationships between collection items.</p><p>In general, collections can be significantly enriched through the addition of events. Events can place collection items into context and create structure by linking related items. This improves the accessibility of the items and helps users to understand their context. Both personal events and historical events are relevant for this cause. Personal events can provide an understanding of the day-today lives of people involved in historical events which can further enable large scale data research. However, detecting and extracting such events is a difficult task, as events are very ambiguous. They lack conclusive definition and their position as events is frequently dependent on personal interpretation.</p><p>In this paper, we create a pipeline that detects, extracts and represents personal and historical events to improve the accessibility of cultural heritage items. We show the added value of enriching the metadata of media objects through events and links between them in the context of DIVE+ <ref type="bibr" target="#b0">[1]</ref>, a Link Data event-centric collections</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">DATASET</head><p>For the current experiment we selected Wikipedia pages related to World War II (WWI), i.e., "Lijsten over de Tweede Wereldoorlog" and "Tweede Wereldoorlog in Nederland". We focused on three themes: deportation, arrest and resistance. This resulted in a corpus of 2.556 pages.</p><p>Person vocabulary: For this research we also used a person vocabulary in order to understand which are the most relevant people in our dataset. The people vocabulary consists of data from Online Begraafplaatsen, Erelijst, and the Oranjehotel Kartotheek, filtered based on our three research themes. Our person vocabulary consists of 35,739 names.</p><p>Personal and Historical Events: The NIOD thesaurus is used as a source of named historical events in our dataset. From the metadata of the person vocabulary aforementioned we identified four personal events: birth, death, arrival at the Oranjehotel and departure from the Oranjehotel.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">EXPERIMENTAL METHODOLOGY</head><p>This section describes our experimental methodology.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Automatic Entity Extraction</head><p>The first step in the pipeline is to detect and extract entities through machines. The input data was processed by two pipelines. The first pipeline uses three free NER tools: DBpediaSpotlight<ref type="foot" target="#foot_1">2</ref> , Targeted Hypernym Discovery<ref type="foot" target="#foot_2">3</ref> (THD) and SemiTags<ref type="foot" target="#foot_3">4</ref> . DBpediaSpotlight extracts DBpedia mentions in texts. The THD extractor uses lexicosyntactic patterns to find hypernyms in the text. The SemiTags extractor is an online tool that recognizes named entities and their meaning in a particular context. As this pipeline is specialized in extracting named entities, it is likely to provide historical named events. The second pipeline that is used is the pipeline of Biogra-phyNet <ref type="bibr" target="#b1">[2]</ref>, a project that uses NLP and Semantic Web Technology to support digital humanities scholars with historical research. The pipeline is a supervised machine earning system trained to extract biographical metadata such as date and place of birth and death, education, occupation, religion, and parents. This pipeline is more likely to provide personal events.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Crowdsourcing experiments</head><p>To improve and enrich the results from the NLP tools described in Section 3.1, we created two crowdsourcing tasks on CrowdFlower. The aim of the first task is to validate and correct entities extracted by the two NLP pipelines and annotate entities that were missed by them. As a proof of concept, for this task we select 100 sentences from the Wikipedia pages. The aim of the second task is to create links between the events and the people, places and time periods identified by the two NLP pipelines and the crowd. For each task we gathered judgments from 15 workers. The task showed one row (sentence) per page and workers were paid 2 cents for one row. The task was only distributed in the Netherlands and Belgium because of language constraints.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Modeling events with Simple Event Model</head><p>To model the events, actors, places and time periods extracted by the machines and the crowd we use the Simple Event Model (SEM). All data in the dataset is being converted to the RDF model of the SEM classes and relationships. We create a first graph by using the output of the second crowdsourcing experiment which is converted to RDF. We create a second graph for the people vocabulary.</p><p>The WWII thesaurus was already available in RDF, in SKOS. This format is also suitable for SEM, and therefore does not have to be converted. Only concepts which are classified as 'gebeurtenis' in the thesaurus are used, plus one manually added named event. To enhance the model, we conduct some alignment proceedings: for the core class place, we automatically generated "sameAs" relationships between identical places in the different graphs. For the core class actor, this can not be done automatically as an identical name does not necessarily address the same person. Therefore, these "sameAs" relationships are done manually.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">RESULTS</head><p>This section describes the results for each step of the experimental methodology.</p><p>Automatic Entity Extraction: Table <ref type="table" target="#tab_0">1</ref> shows the results of the two pipelines. The THD extractor extracted 145,617 unique entities, mostly locations months, out of which 6,675 entities are events. However, when analyzing the results from THD there is a lot of noise, e.g., "handen namen" is classified as the Belgian city "Namen". DBpediaSpotlight extracted 115,984 unique entities, with a total number of 2,518 entity types (34,785 entities were extracted without type). While performing well on recognizing locations and persons, DBpediaSpotlight also extracted 3,434 events. The SemiTags extractor extracted 99,170 unique entities, but no entity was classified as event. With the BiographyNet pipeline we extracted a number of 260,244 unique entities. The BiographyNet pipeline is especially good at recognizing events, names and precise dates.</p><p>Crowdsourcing Results: The 100 sentences ran through the crowdsourcing pipeline had 431 entities automatically extracted: 132 events, 74 dates, 128 locations and 97 actors. In the first crowdsourcing task the workers annotated 167 extra entities and removed 34 entities. This resulted in a total of 564 entities, consisting of 102 dates, 228 events, 125 places and 61 actors. For the second crowdsourcing task we used 71 sentences to connect events and places, 72 sentences to connect events and actors and 77 sentences to connect events and time periods. From these sentences, a total of 94 events were linked to 58 places, 61 actors and 64 time periods.</p><p>Event Modeling: After we converted all the data into SEM classes we obtained a knowledge graph with 209,183 triples. After generating "owl:sameAs" relations between names and locations, the knowledge graph extended to 209,245 triples. Looking at the four core classes in SEM, our enrichment contains 18,316 personal and historical event triples, 9,300 actor triples, 875 place triples and 60 time periods, as shown in Table <ref type="table" target="#tab_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">ILLUSTRATIVE EXAMPLE</head><p>To illustrate the effectiveness of SEM for our dataset, this section describes an illustrating example through the interface of the DIVE+ demonstrator. When searching for the historical event "februaristaking" in DIVE+ all related events, objects, locations and actors are shown (see Figure <ref type="figure" target="#fig_0">1</ref>). When looking at one of the objects we see all related entities to that specific object. For example, we can further explore the entity Communistische Partij Nederland (CPN), Figure <ref type="figure" target="#fig_1">2</ref>, which is an important concept for the February Strike. As objects are linked to each other we make the objects easier accessible and create meaningful relations between them. With the use of the event model we tell stories around and about the object and the scholars can get a more complete overview of the event. For the personal event, we focus on Eduard Popko van Groningen. Without the model, information about him is fragmented over the collections of Erelijst and Online Begraafplaatsen. All this information is now available in DIVE+ through one query. Figure <ref type="figure" target="#fig_2">3</ref> shows that the query leads to an actor, which is related to five events. The first four are the events of birth and death as described  in their original source. The fifth event is an event extracted from a Wikipedia page. The related entities are, in turn, also related to other entities. In the case of the extracted event, to the actor, Eduard, and a location, Sachsenhausen. New opportunities for digital humanities scholars appear through the entity 'Sachsenhausen'. The location Sachsenhausen is linked to 21 events which are all connected with other actors. This means that potential fellow prisoners of Eduard are known. Another storyline could be that the actors of the other 21 events were also shot at the 3rd of May 1942, which would indicate that a bigger event took place.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">CONCLUSION AND FUTURE WORK</head><p>Digitization projects made various cultural heritage collections publicly available online. However, due to unstructured metadata and the distributed allocation of the objects in different databases, purely digitizing objects does not make them more accessible. In this research we used a hybrid machine-crowd approach to extract personal and historical events and use them for the enrichment of </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Media Objects, Events, Actors and Places related to 'Februaristaking'</figDesc><graphic coords="2,329.63,422.04,216.89,173.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Exploration of related concept 'CPN'</figDesc><graphic coords="3,65.48,277.76,216.88,188.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Events related to 'Eduard Popko van Groningen'</figDesc><graphic coords="3,329.63,187.70,216.88,151.82" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Overview of Entities Extracted by each Tool</figDesc><table><row><cell>Tool</cell><cell>#Pages Processed</cell><cell>#Entities Extracted</cell><cell>% of Unique Entities</cell><cell>Avg. # Entities per Page</cell><cell>#Entity Types</cell><cell># Entities without Type</cell></row><row><cell>THD</cell><cell>2,665</cell><cell>145,617</cell><cell>73.50%</cell><cell>64.64</cell><cell>2,517</cell><cell>0</cell></row><row><cell>DBpediaSpotlight</cell><cell>2,665</cell><cell>115,984</cell><cell>37.69%</cell><cell>43.52</cell><cell>193</cell><cell>34,785</cell></row><row><cell>SemiTags</cell><cell>2,665</cell><cell>99,170</cell><cell>31.37%</cell><cell>37.21</cell><cell>4</cell><cell>0</cell></row><row><cell>BiographyNet</cell><cell>2,308</cell><cell>349,321</cell><cell>32.62%</cell><cell>151.35</cell><cell>11</cell><cell>0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Triples Overview per SEM Core Classes</figDesc><table><row><cell cols="2">SEM core class Triples</cell></row><row><cell>Event</cell><cell>18,317</cell></row><row><cell>Actor</cell><cell>9,300</cell></row><row><cell>Place</cell><cell>875</cell></row><row><cell>Time</cell><cell>60</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://www.niod.nl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://www.dbpedia-spotlight.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">http://entityclassifier.eu/thd/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">http://nlp.vse.cz/SemiTags/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We used automatic event extraction, making use of different information extracting tools and manual event curration through the use of a crowdsourcing strategy. Applied our methodology a set of Wikipedia pages related to WWII. We illustrated the historical and personal events in the DIVE+ demonstrator. With our created historical and personal event model we are able to enrich the metadata of the objects in the NOB portal. In this way the available digital resources of the NOB portal can be linked and improved with storylines told about and around the objects.</p><p>Future work can improve the pipeline. For example, the entity extraction could be improved or tested with other NLP tools better suited for Dutch texts. Another way to improve entity extraction is by tweaking the crowdsourcing tasks. Referencing words, such as 'he' and 'that' were not extracted as the context around the sentence is missing. As a consequence, a lot of verbs could not be linked with an actor. We recommend doing a pilot with extracting entities from paragraphs to improve results. Lastly, we suggest enriching the SEM with adding subtypes to the core classes actor, event, place and time. For example, nationalities can be added to names of people for a richer context.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>SEMANTiCS 2017 workshop proceedings: EVENTS September 11-14, 2017, Amsterdam, Netherlands browser. This research is done on behalf of the Institute for War, Holocaust and Genocide Studies 1 (NIOD).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">DIVE into the event-based browsing of linked historical media</title>
		<author>
			<persName><forename type="first">Victor</forename><surname>De</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Boer</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Web Semantics: Science, Services and Agents on WWW</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="152" to="158" />
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">BiographyNet: Methodological Issues when NLP supports historical research</title>
		<author>
			<persName><forename type="first">Antske</forename><surname>Fokkens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Serge</forename><surname>Ter Braake</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Niels</forename><surname>Ockeloen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piek</forename><surname>Vossen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Susan</forename><surname>Legêne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guus</forename><surname>Schreiber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LREC</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Improving the geospatial consistency of digital libraries metadata</title>
		<author>
			<persName><forename type="first">Walter</forename><surname>Renteria-Agualimpia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Francisco</forename><forename type="middle">J</forename><surname>Lopez-Pellicer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Javier</forename><surname>Lacasta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Javier Zarazaga-Soria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pedro</forename><forename type="middle">R</forename><surname>Muro-Medrano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="507" to="523" />
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
