<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Experiences of Using WDumper to Create Topical Subsets from Wikidata</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
				<date type="published" when="2021-02-04">4 February 2021</date>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Seyed</forename><surname>Amir</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Hosseini</forename><surname>Beghaeiraveri</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Alasdair</forename><forename type="middle">J G</forename><surname>Gray</surname></persName>
						</author>
						<author role="corresp">
							<persName><forename type="first">Fiona</forename><forename type="middle">J</forename><surname>Mcneill</surname></persName>
							<email>f.j.mcneill@ed.ac.uk</email>
							<affiliation key="aff1">
								<orgName type="department">School of Informatics</orgName>
								<orgName type="institution">The University of Edinburgh</orgName>
								<address>
									<settlement>Edinburgh</settlement>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department">School of Mathematical and Computer Sciences</orgName>
								<orgName type="institution">Heriot-Watt University</orgName>
								<address>
									<settlement>Edinburgh</settlement>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Experiences of Using WDumper to Create Topical Subsets from Wikidata</title>
					</analytic>
					<monogr>
						<imprint>
							<date type="published" when="2021-02-04">4 February 2021</date>
						</imprint>
					</monogr>
					<idno type="MD5">C8694560B371CD8E9620E6C55F7112B3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T02:40+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>wikidata</term>
					<term>knowledge graph subsetting</term>
					<term>topical subset</term>
					<term>wdumper</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Wikidata is a general-purpose knowledge graph covering a wide variety of topics with content being crowd-sourced through an open wiki. There are now over 90M interrelated data items in Wikidata which are accessible through a public query endpoint and data dumps. However, execution timeout limits and the size of data dumps make it difficult to use the data. The creation of arbitrary topical subsets of Wikidata, where only the relevant data is kept, would enable reuse of that data with the benefits of cost reduction, ease of access, and flexibility. In this paper, we provide a working definition for topical subsets over the Wikidata Knowledge Graph and evaluate a third-party tool (WDumper) to extract these topical subsets from Wikidata.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>A Knowledge Graph (KG) is defined as representing real-world entities as nodes in a graph with the relationships between them captured as edges <ref type="bibr" target="#b9">[10]</ref>. In recent years, there have been a growing number of publicly available KGs; ranging from focused topic specific ones such as GeoLinkedData <ref type="bibr" target="#b11">[12]</ref>, EventMedia <ref type="bibr" target="#b10">[11]</ref>, and UniProt <ref type="bibr" target="#b15">[16]</ref>, to more general knowledge ones such as Freebase <ref type="bibr" target="#b5">[6]</ref>, DBpedia <ref type="bibr" target="#b2">[3]</ref>, and Wikidata <ref type="bibr" target="#b16">[17]</ref>. These general purpose KGs cover a variety of topics from sports to geography, literature to life science, with varying degrees of granularity.</p><p>Wikidata <ref type="bibr" target="#b16">[17]</ref> is a collaborative open KG created by the Wikimedia Foundation. The main purpose of Wikidata is to provide reliable structured data to feed other Wikimedia projects such as Wikipedia, and is actively used in over 800 Wikimedia projects 3 . It contains over 90 million data items covering over 75 thousand topics <ref type="foot" target="#foot_0">4</ref> . Regular dumps of the data are published in JSON and RDF under the Creative Commons CC0 public license. However, the size of the gzipped download files has grown from 3GB in 2015 to 85GB in 2020, and keeps increasing as more data is added. The size of these files make it increasingly difficult and costly for others to download and reuse, particularly if only focused on a particular topic within the data, e.g. life sciences data or politicians. While Wikidata can be queried directly through an open SPARQL endpoint <ref type="foot" target="#foot_1">5</ref> , it is subject to usage limits (as are other public endpoints of KGs) which limits the scale and complexity of queries that can be issued <ref type="bibr" target="#b16">[17]</ref>. Thus there is a need for topical subsets of large KGs that contain all data on a specific topic to enable complex analysis queries to be conducted in a cost effective and time efficient way.</p><p>Our motivation for research on Wikidata subsetting is a combination of research goals, flexibility and ease of use. From the flexibility and ease of use perspective, we are looking for Wikidata subsets that can allow users to run smaller versions of Wikidata on available platforms such as laptops and PCs. Wikidata, as a knowledge graph with an interesting data model, has significant features for inspiration and improvement, but the speed of research and the diversity of researchers will be reduced if any experiment on it requires powerful servers, processing clusters, and hard disk arrays. From the research point of view, our motivation is creating a type of subset we call Topical Subset which is a set of entities related to a particular topic and the relationships between them. Having topical subsets of Wikidata for example in the fields of art, life sciences, or sports, not only helps us achieve the first goal (flexibility and ease of use) but also provides a platform for comparing and evaluating Wikidata features in different topics. The subset will also make the research experiments reproducible as the data used can be archived and shared more easily.</p><p>You can envision that such subsets can be generated through SPARQL CONSTRUCT queries. While this is straightforward for small subsets focused on a single entity type, e.g. politicians, it does not scale to interrelated topics that make up a larger domain, e.g. the life sciences subset defined by GeneWiki <ref type="bibr" target="#b17">[18]</ref>. In this paper, we present our experiences of defining and creating topical subsets over Wikidata using WDumper <ref type="bibr" target="#b0">[1]</ref>; a third-party tool provided by Wikidata to create custom RDF dumps of Wikidata. We define topical subsets over a KG (Section 3.2) and evaluate WDumper as a practical tool to extract such subsets (Section 4).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>KG subsetting, particularly in the context of Wikidata <ref type="bibr" target="#b1">[2]</ref>, has been gaining attention in the last couple of years, with use cases and potential approaches being explored <ref type="bibr" target="#b6">[7]</ref>. Although there have been several approaches to subsetting proposed, to the best of our knowledge there is no agreed definition for a topical subset nor a unified and evaluated way to create such subsets. The Graph to Graph Mapping Language (G2GML) enables the conversion of RDF graphs into property graphs using an RDF file or SPARQL endpoint and a mapping file <ref type="bibr" target="#b12">[13]</ref>. This could be exploited to create a topical subset based on the definition in the mapping file. However the output would be a property graph. Context Graph <ref type="bibr" target="#b13">[14]</ref> can be used to generate subsets of an RDF KG, and was developed to create subsets for testing knowledge base completion techniques. The approach captures all nodes and edges within a given radius of a set of seed nodes. While the generated subsets are suitable for testing knowledge base completion techniques, there is no guarantee to the topical cohesion of the subset, and thus do not meet the needs identified for this work. YAGO4 <ref type="bibr" target="#b14">[15]</ref> is a collection of Wikidata instances under the type hierarchy of schema.org <ref type="bibr" target="#b8">[9]</ref>. It suggests a new logically consistent knowledge base over the Wikidata A-Boxes, however, there is no choice as to what topic be in the final KG.</p><p>WDumper <ref type="bibr" target="#b0">[1]</ref> is a third-party tool for creating custom and partial RDF dumps of Wikidata suggested at the Wikidata database download page <ref type="foot" target="#foot_2">6</ref> . The WDumper backend uses the Wikidata Toolkit (WDTK) Java library to apply filters on the Wikidata entities and statements, based on a specified configuration that is created by its Python frontend. This tool needs a complete JSON dump of Wikidata and creates an N-Triple file as output based on filters that the config file explains. The tool can be used to create custom subsets of Wikidata. In Section 4.5 we will investigate whether it can generate topical subsets.</p><p>Shape Expressions (ShEx) <ref type="bibr" target="#b7">[8]</ref> is a structural schema language allowing validation, traversal and transformation of RDF graphs. By caching triples that match the constraints, referred to as "slurping", a subset of the dataset can be created that conforms to the ShEx schema. Therefore to create a topical subset, all that is required is the definition of the ShEx schema. However, this approach is not available at scale yet and thus cannot be applied to Wikidata.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Knowledge Graph Subsetting</head><p>General purpose KGs like Wikidata are valuable sources of facts on a wide variety of topics. However, their increasing size makes them costly and slow to use locally. Additionally, the large volume of data in Wikidata increases the time required to run complex queries. This often restricts the types of queries that can be posed over the public endpoint since it has a strict 60-second limit on the execution time of queries. Downloading and using a local version of Wikidata is a way of circumventing the timeout limit. However, this is not a cheap option due to the size of the data. A suggested system to have a personal copy of Wikidata includes 16 vCPUs, 128GB memory, and 800GB of raided SSD space <ref type="foot" target="#foot_3">7</ref> .</p><p>There are a large number of use case scenarios where users will not need access to all topics in a large general purpose KG. A small and complete enough subset can be more suitable for many purposes. With a small subset inference strategies can be applied to the data and complete in reasonable time. Topical subsets could also be published along with papers, which provides better reproducibility of experiments. Therefore having a topical subset that is smaller but has the required data can enable complex query processing on cheap servers or personal computers -reducing the overall cost -whilst also providing an improvement in query execution times.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Topical Subset Use Cases</head><p>We now define four use cases for topical subsets in Wikidata that we will use to review WDumper, and can also be used in other reviews as a comparison platform. Note that the use cases are defined in terms of English language statements. A subsetting approach, method, or tool would need to formalise these, as appropriate for their configuration, to extract the relevant data.</p><p>Politicians: This subset should contain all entities that are an instance of the class politician, or any of its subclasses. In the case of Wikidata, this would be the class Q82955, while for DBpedia it would be the class Politician. The subset should contain all facts pertaining to these entities, i.e. in Wikidata all statements and properties. General(military) Politicians: The subset should contain all entities that are an instance of the class politician(Q82955) or any of its subclasses, who also are a military officer (Q189290) and have the rank of general (Q83460), i.e. politico-military individuals. The main goal of this use case is to see the effect of having more conditions in the English definition on the run-time and the volume of the output of subset extraction tools. UK Universities: The subset should contain all instances of the class university(Q3918) or any of its subclasses, that are located in the UK. The subset should contain all statements and properties pertaining to these entities. This use case extends the complexity of the subset by having alternative properties and values to satisfy, e.g. the location can be captured in Wikidata with the properties country(P17), located in territory(P131), or location(P276).</p><p>Likewise the country could be stated as one of the component parts of the UK, e.g. Scotland. Gene Wiki: This case is based on the class-level diagram of the Wikidata knowledge graph for biomedical entities given in <ref type="bibr" target="#b17">[18]</ref>. The class-level diagram specifies 17 different item types from Wikidata mentioned in the Gene Wiki project. The subset should contain all instances of these 17 classes and their subsets.</p><p>The selection of these use cases is a combination of research and experimental goals. The Gene Wiki and Politicians use cases have been selected for future research purposes because of their hypothetical richness in references. The other Listing 1.1. An example of a function R which is a query to return all entities with type city SELECT ? entity WHERE { ? entity wdt : P31 wd : Q515 . # instance of ( P31 ) city ( Q515 ) } two use cases have been chosen to explore the expressiveness of a topical subset definition, and then to explore the runtime execution of these.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Topical Subset Definition</head><p>We now provide a definition for topical subsets based on the Wikidata data model. Wikidata consists of the following collections:</p><p>-E: set of Wikidata entities -their ID starts with a Q.</p><p>-P : set of Wikidata properties -their ID starts with a P.</p><p>-S: set of Wikidata statements. Now we define the filter function R : E → E as a black-box that can be applied on E and selects a finite number of its members related to a specific topic. Let E R ⊆ E be the output of the function R. For entity e ∈ E let S e ⊆ S be all simple and complex Wikidata statements in which e is the subject. Note that in Wikidata, a simple statement is a regular RDF triple, while a complex statement is a triple that references and/or qualifiers attached to it. Also, let P e be all properties which are used in S e triples either for the statement itself or qualifiers/references. With these assumptions, we define dump D R as a topical subset of Wikidata with respect to R:</p><formula xml:id="formula_0">D R := (E R , e∈E R P e , e∈E R S e )</formula><p>From the definitions of P e and S e we can conclude that e∈E R P e ⊆ P and e∈E R S e ⊆ S and subsequently D R is a subset of Wikidata. We consider R as black-box; the input of R is the set of all Wikidata entities and its output is a subset of Wikidata entities related to a specific topic. The function R can be any set of definitions, rules, or filters that describe a related group of entities. The definition of R depends on the topic that is being described. One example of R is a simple SELECT query that describes all entities that have type city(Q515) (Listing 1.1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">WDumper</head><p>WDumper is a tool provided by Wikidata for producing custom subsets from Wikidata<ref type="foot" target="#foot_4">8</ref> . The R function can be seen as a filter approach on entities. For each topic, the appropriate filters on entities must be defined. Once the filters are defined and the subset E R is extracted, WDumper extracts all statements with origin e, where e ∈ E R , along with their qualifiers and references. A component overview of WDumper is given in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>WDumper requires two inputs. The first is the complete dump of the Wikidata in JSON. The second is a JSON specification file that contains rules and filters for determining which entities, properties and statements to extract from the full Wikidata dump. This is the definition of the function R. The output of WDumper is an N-Triple (.nt) file that contains the entities and statements specified in the second input. There is also a GUI for creating the input specification file. We review WDumper through the following steps:</p><p>1. Writing WDumper specifications for the use cases in Section 3.1. 2. Running WDumper with the above specifications on two complete Wikidata dumps belonging to two different time points and compare the run-time and the volume of the extracted output. 3. Evaluating the extracted output via performing different queries both on the output and the input full dump. 4. Summarizing results and expressing strengths and weaknesses of WDumper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">WDumper Specification Files</head><p>Section 3.1 introduced use cases to evaluate subset extraction tools. In this section, we describe the corresponding WDumper specification files <ref type="bibr" target="#b4">[5]</ref> generated using the GUI depicted in Figure <ref type="figure" target="#fig_1">2</ref>. The GUI provides several controls as to what to include in the subset. The first is to select whether you are interested in items or properties. Using properties allows you to extract a subset of the Wikidata ontology, while items returns data instances. Items can be further filtered by giving a property and value. Other options then permit you to select whether all statements are returned ("any") or just the top ranked statements ("best rank") <ref type="foot" target="#foot_6">9</ref> . These filters allow the WDumper to extract the intermediate nodes of statements, references, and qualifiers. In all use cases, we created a specification with and without the additional references and qualifiers. This allows us to investigate the effect of statement filters on execution time and output volume. In the specification file, these filters can be seen in the statement subarray, as "qualifiers", "references", "simple", "full" keys which are false or true respectively.</p><p>Politicians: We define a filter on the occupation property(P106) of the entities to be a politician(Q82955). General(military) Politicians: This extends the Politicians definition with two more conditions. The first is the occupation property(P106) to be military officer(Q189290). The second is the military rank property(P410) to be general(Q83460). UK Universities: We define a filter on entities with two conditions: the instance of property(P31) to be university(Q189290), and the country property(P17) to be United Kingdom(Q145). Gene Wiki: For each item type of the class-diagram in <ref type="bibr" target="#b17">[18]</ref>, we create a filter on the corresponding entities in WDumper via the instance of property(P31) to be gene(Q7187), protein(Q8054), etc. In this case, no filters are defined on the types as we require all statements associated with the types ot be in the subset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Experimental Setup</head><p>We now give details of our experimental environment in which we will evaluate WDumper.</p><p>Input Dumps. We use two full dumps of the Wikidata database. The first full dump is from 27 April 2015<ref type="foot" target="#foot_7">10</ref> , and the second is from 13 November 2020 <ref type="foot" target="#foot_8">11</ref> . The selected 2020 dump was the latest JSON dump available when conducting our evaluation. The selected 2015 dump is the first archive date for which both JSON and Turtle files are available (we need the JSON file for WDumper running, while the Turtle file is needed to import the full dump in a triplestore and evaluate output of WDumper based on the input). Experimental Run. The calculated times have been extracted from the elapsed time mentioned in WDumper output log. For each of the execution cases, three independent runs were performed. The average and standard deviation from these times were calculated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Evaluating WDumper</head><p>WDumper was run with the specification files described in Section 4.1 and the two Wikidata dumps. The generated subsets are included in <ref type="bibr" target="#b4">[5]</ref> and the results stated in Table <ref type="table">2</ref>. For each use case, we generated subsets of simple statements and with the inclusion of statement nodes, references, and qualifiers (labelled "withRQFS" in the table <ref type="table">)</ref>. Initial observations from Table <ref type="table">2</ref> show that the run-time on the 2020 dump is significantly longer than the 2015 dump. This can be justified by the larger volume data that must be processed to produce the subset. In all cases, generating the subset with additional statements took longer than generating the simple statements, and produced more volume in the output, indicating the addition of references, qualifiers, and statements nodes in the output. For example, this change is very significant in the case of Gene Wiki in the 2020 dump. The added filters as well as the conditions added to the filters also have a direct effect on the run-time and an inverse effect on the output volume, which is to be expected. Of course within run-times, the amount of data that must be written in the output must also be considered. This is evident in comparisons between UK universities and the military politicians in which the volume of data has a greater impact than the number of conditions. Overall, considering the high volume of data, the time required to extract a topical subset by WDumper seems appropriate.</p><p>Table <ref type="table">2</ref>. Time elapsed and the size of the output in running WDumper on each use case and full dumps. "withRQFS" label denotes the specification that aims to extract statement nodes, references, and qualifiers. The FoE column denotes the number of filters on entities. The CiF column denotes the number of conditions in each filter. The average of three runs are reported (AVG column) together with the standard deviation (SD column). Sizes belong to compressed nt.gz files that are the direct output of WDumper. Inside parentheses, sizes and times are converted to other units for more readability (h for hours, KB for kilobytes, and MB for megabytes). Adding more filters does not have a huge effect on runtime which is dominated by data volume.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Use case</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Topical Subset Validation</head><p>The previous section considered the runtime performance of the WDumper tool and the size of the generated subsets. We now consider validating the content of the subsets. That is, we consider if the produced output has the information that it was supposed to have according to the definition in Section 3.2. Our assessment will be based on the following conditions:</p><p>Condition 1: The number of filtered entities in the output should be equal to the same number of entities in the input dump. For example, in the Politicians use case, the number of persons with the occupation of politician in output should be equal to the number of persons with the occupation of politician in the corresponding input dump. This condition can be tested with COUNT queries on the input and output datasets. and qualifiers exactly as they are at the input dump. This condition can be tested by querying the qualifiers and references of some given statements.</p><p>Testing the conditions requires running different queries on the input and output of WDumper. For the output of WDumper, we use Apache Jena Fuseki version 3.17 to import data as TDB2 RDF datasets and perform queries.</p><p>Data Corrections. We encountered problems loading the 2015 turtle dump<ref type="foot" target="#foot_10">14</ref> into Fuseki. Errors arose from the inclusion of bad line endings in more than 100 cases, and the existence of characters such as '\a'. Unacceptable characters such as '\n' and '\\' can also be seen in the WDumper outputs, which reinforces the possibility that this problem occurs due to the conversion of information from the JSON file to RDF format.</p><p>For the WDumper outputs and the 2015 dump, these errors were manually fixed using the sed commands given in Listing 1.2. The sanitized versions are available in <ref type="bibr" target="#b3">[4]</ref>. In the case of 2020 dump, we use Wikidata Query Service (WDQS) due to the time that would be required to fix and load this data into Fuseki. The date of implementating our evaluation queries is approximately two months after the creation date of the 2020 dump (27 November 2020). In this period, new data may have entered Wikidata which are available by WDQS and are not present in 2020 dump (and subsequently are not present in the WDumper output). Because of this, there may be slight differences in the counts of entities and statements between input and output that is not related to WDumper functionality. We tried to use Wikidata history query service <ref type="foot" target="#foot_11">15</ref> to quantify the rate of Wikidata increases in this period, but the history covers a range from the creation of Wikidata to July 1st 2019.</p><p>Validation of Condition 1. We use COUNT queries to validate this condition. The purpose of these queries is to count the entities that should be in the output according to the filter(s) of each use case. If WDumper is performing correctly, Listing 1.3. COUNT queries for evaluating condition 1. Prefixes and most of Gene Wiki's query have been deleted for more readability.  the result of this count should be the same on both the input and output datasets. These queries will be different for each use case, depending on the definition of that use case. For example, while in the Politicians use case we count the number of people with political jobs, in the case of Gene Wiki we count the union of entities of type disease, genes, proteins, etc. Listing 1.3 shows the queries executed for each use case. These queries run on the each use case's output "withRQFS" since these are included in the input dataset. The results of performing the COUNT queries are shown in Table <ref type="table" target="#tab_3">3</ref>.</p><formula xml:id="formula_1">############ Politicians # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # SELECT ( COUNT ( DISTINCT ? item ) AS ? count ) WHERE { ? item wdt : P106 wd : Q82955 . # occupation of politician } ############ UK Universities # # # # # # # # # # # # # # # # # # # # # # # # # # SELECT (</formula><formula xml:id="formula_2">############ Gene Wiki # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # SELECT (</formula><p>Our results show that for the 2015 dump, the number of entities in the output and input is equal except for the Politicians use case. In both 2015 and 2020 dumps, the difference between input and output is less than one percent in the cases of inequality. In the case of 2020 dump, the difference can be attributed to the entry of new data in the interval between our tests and the dump date. This is reasonable especially in the case of Gene Wiki where bots are importing new information into Wikidata every day. In the case of the 2015 dump in the Politicians row, the 35 differences between input and output is unjustifiable. The reason for this difference may be the inability of WDumper to parse the data of these entities in the input dump. WDumper uses the JSON file as input, and to be able to fetch an entity, it must see the specific structure of the Wikidata arrays and sub-arrays in the JSON file. Some entities may not have this complete structure in the JSON file but they do exist in the Turtle file.</p><p>Validation of Condition 2. To validate this condition, in each use case we use DESCRIBE queries for an arbitrary entity that is in the WDumper output. DESCRIBE queries list all triples of the given entity. We expect that the result of the DESCRIBE queries should be the same on both the input and output datasets. For each use case, we selected an arbitrary entity (called Tested Entity) which is present in both the input and output dataset. We then run a DESCRIBE wd:Q... query and count the extracted triples. Table <ref type="table" target="#tab_4">4</ref> shows our results.</p><p>From Table <ref type="table" target="#tab_4">4</ref> it is clear that the number of triples in the DESCRIBE queries in both the 2020 and 2015 dumps are not equal. This difference prompted us to explore the differences using the compare module of the RDFlib library. It was found that in the case of the 2015 dump, the input dump contains predicates such as &lt;http://www.w3.org/2004/02/skos/core#prefLabel&gt; and &lt;http://schema.org/name&gt;, which are not extracted by WDumper. Table <ref type="table" target="#tab_5">5</ref>, shows the details and total numbers of predicates that are in the input dump (2015 dump) for the selected entities and WDumper could not fetch. As we can see, the total column is exactly the difference between the describe queries. In the case of 2020 dump, some of predicates with &lt;http://www.w3.org/2004/02/ skos/core#&gt; prefix, such as dateModified, and all &lt;http://www.wikidata. org/prop/direct-normalized/&gt; predicates are not detectable by WDumper. However, in both dumps the statements whose predicate is a property of Wikidata (e.g. P31, P106, etc.), were completely extracted by WDumper. Validation of Condition 3. To validate this condition, we selected an arbitrary entity from each use case, and for this entity, we considered one of its statements.</p><p>We then counted the qualifiers and the references of this statement in 2020 dump (over the WDQS) and in the output of WDumper. Table <ref type="table" target="#tab_6">6</ref> shows the selected entity, selected property, and the number of qualifiers and references for them.</p><p>From Table <ref type="table" target="#tab_6">6</ref>, it is clear that WDumper can extract qualifiers and references completely from the input.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Summarizing Results, Strengths and Weaknesses</head><p>The results of our evaluations show that WDumper, as a custom dump production tool, can be used to create some topical subsets. This tool can correctly and completely extract the entities specified by its filters. It also extracts almost all statements related to entities (except it is not designed to extract some prefixes). One of the features we have been looking for is the ability to extract references and qualifiers of Wikidata statements, which WDumper can do. Setting up this tool is not very complicated; the user only needs to select the filters of the entities and statements, run the tool and it extracts all of the information at once. Its GUI is also somewhat helpful, while the JSON structure of its specification files is also simple and understandable.</p><p>Limitations and Weaknesses. The most important weakness of WDumper with regard to the topical subsets is the limitation in the definition of entity filters. In WDumper, entities can only be filtered based on the presence of a P x property or having the value v for a P x property. Although it is possible to deploy any number of such filters, this is not enough to specify some kinds of use cases.</p><p>For example, suppose we want to specify the Scottish universities subset. By reviewing some of these universities on the Wikidata website, we can find that their corresponding entity does not have any property that directly indicates they belong to Scotland. Of course, we can define the R function of these subsets through indirect methods (for example, considering the Geo-location of entities in Scottish universities), but these type of filters are not available in WDumper. The recognition of type hierarchies is another limitation of WDumper. In the case of UK universities, for example, the University of Edinburgh(Q160302) is not among the universities extracted by WDumper. The reason for this is that the instance of property(P31) in this university refers to public university(Q875538) instead of university(Q3918). In SPARQL queries, such cases are handled by the property paths like wdt:P31/wdt:P279 * . These property paths are not available in WDumper. The strategy of considering more filters to cover all subtypes needs a comprehensive knowledge of Wikidata ontology. This strategy will fail if the class hierarchy changes.</p><p>Another limitation is the inability to communicate between different filters in multi-filter cases. For example, in the Gene Wiki use case, we may want diseases that are somehow related to a gene or protein, while in the WDumper output there are diseases that have nothing to do with genes, proteins, and other Gene Wiki types. The inability to choose another output format other than N-Triples, especially the Wikibase JSON output, which is more suitable for using the subset produced in a Wikibase instance and also has a smaller volume, is another limitation.</p><p>The main implications of these limitations is the reduction of flexibility of subset extracting with this tool. With these weaknesses, users have to spend much more time defining the desired subset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>In this paper, we reviewed the issue of building topical subsets over Wikidata. Our motivation for topical subsets is to enable efficient evaluation of complex queries over the knowledge graph with lower costs, reproducibility of experiments through archiving datasets, ease of use, and flexibility. We provided example use cases for topical subsets as well as a definition for topical subsets. This definition enables us to evaluate and compare subset creation tools.</p><p>In this study we used WDumper for topical subset extraction over Wikidata and tested it by measuring run-time and output volume on four different use cases. We evaluated the correctness of the subsets generated by WDumper by comparing the answers to queries over the subsets and the full knowledge graph. Our experience shows that WDumper can be used to generate topical subsets of Wikidata in some use cases but not for all use cases. WDumper can extract the entities specified by its filters and extract most statements related to those entities; it also fetches the statement nodes and references/qualifiers. However, WDumper has some weaknesses regarding topical subsets. Its main problem is the way it defines filters on entities that reduces the power of this tool to build topical subsets. The most tangible issue is the inability to define and fetch subclasses of a class of entities, which is important in many use cases. Our suggestion for the future works is to explore alternative subsetting approaches such as using SPARQL queries or Shape Expressions. With selectors like SPARQL queries or ShEx schemata, we can increase the expressivity of the subset creation. It will also allow for subsets to be created on Knowledge Graphs other than Wikidata.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Component overview of WDumper</figDesc><graphic coords="6,186.64,115.83,242.08,94.08" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. WDumper Configuration GUI for the politicians use case.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>COUNT ( DISTINCT ? item ) AS ? count ) WHERE { ? item wdt : P31 wd : Q3918 ; # instance of university wdt : P17 wd : Q145 . # country of United Kingdom }############ General ( military ) Politicians ############ SELECT (COUNT ( DISTINCT ? item ) AS ? count ) WHERE { ? item wdt : P106 wd : Q82955 ; # occupation of politician wdt : P106 wd : Q189290 ; # occupation of milit . officer wdt : P410 wd : Q83460 . # military rank of general }</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Table 1 provides summary information Details of the Wikidata dumps. The Total Items and Total Statements columns are obtained from the Wikidata Stats tool 13 . The 2015 dump is smaller, can be stored and processed locally even on PCs, and it takes a much shorter time to generate output. For this reason, it is very suitable for initial tests. The 2020 dump, on the other hand, is much richer and can provide insights on how WDumper deals with large datasets of the size that Wikidata now produces. Experimental Environment. Experiments were performed on a multi-core server with 64 8-core AMD 6380 2.5GHz 64bit CPUs, 512GB of memory and 2.7TB disk space. Java openJDK version 11 (build 11+28) and Gradle 6.7.1 was used to compile and run WDumper.</figDesc><table><row><cell cols="4">Release Date JSON Compressed Size (GB) Total Items Total Statements</cell></row><row><cell>2015-04-27</cell><cell>4.52</cell><cell>17,632,496</cell><cell>61,427,934</cell></row><row><cell>2020-11-13</cell><cell>90.42</cell><cell>90,368,519</cell><cell>1,153,949,899</cell></row><row><cell>about these two dumps.</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>Listing 1.2. Commands used for fixing syntax errors of the 2015 dump. For each entity that is supposed to be in output, the number of its related statements must be equal in both input and output datasets. For example, in the Politicians use case, if the main dump has 50 statements about George Washington, we expect to see the same number of statements about this politician in the output too. This condition can be tested using DESCRIBE queries. Condition 3: WDumper can extract intermediate statement nodes, references,</figDesc><table><row><cell>sed -i -E 's /( &lt;.*)}(.* &gt;)/\1\2/ ' &lt; dump_file &gt;</cell></row><row><cell>sed -i -E 's /( &lt;.*)\\ n (.* &gt;)/\1\2/ ' &lt; dump_file &gt;</cell></row><row><cell>sed -i -E 's /( &lt;.*)\|(.* &gt;)/\1\2/ ' &lt; dump_file &gt;</cell></row><row><cell>Condition 2:</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 .</head><label>3</label><figDesc>Results of performing COUNT queries of each use case (Listing 1.3) on the output of WDumper and input full dump for both 2015 and 2020 dumps. The last column are COUNT results queried against WDQS instead of the 2020 dump itself.</figDesc><table><row><cell cols="3">COUNT ( DISTINCT ? item ) AS ? count ) WHERE {</cell><cell></cell><cell></cell></row><row><cell>{? item wdt : P31 wd : Q423026 .}</cell><cell cols="3"># instance of active site</cell><cell></cell></row><row><cell>UNION</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">{? item wdt : P31 wd : Q4936952 .} # instance of anat . struct .</cell><cell></cell></row><row><cell>UNION</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell># ...</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>UNION</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">{? item wdt : P31 wd : Q50379781 .} # instance of therap . use</cell><cell></cell></row><row><cell>}</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Use case</cell><cell></cell><cell cols="2">2015 Dump Output Input</cell><cell cols="2">2020 Dump Output Input</cell></row><row><cell>Politicians</cell><cell></cell><cell>246,009</cell><cell>246,044</cell><cell>641,387</cell><cell>646,401</cell></row><row><cell cols="2">General (military) Politicians</cell><cell>165</cell><cell>165</cell><cell>597</cell><cell>602</cell></row><row><cell>UK Universities</cell><cell></cell><cell>73</cell><cell>73</cell><cell>183</cell><cell>186</cell></row><row><cell>Gene Wiki</cell><cell></cell><cell>19,432</cell><cell cols="3">19,432 3,282,560 3,283,471</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 .</head><label>4</label><figDesc>Results of performing DESCRIBE queries on the selected entity.</figDesc><table><row><cell>Use case</cell><cell>Entity</cell><cell cols="4">2015 Dump 2020 Dump Output Input Output Input</cell></row><row><cell>Politicians</cell><cell>Q23</cell><cell>408</cell><cell>776</cell><cell>871</cell><cell>921</cell></row><row><cell>General (military) Politicians</cell><cell>Q355643</cell><cell>104</cell><cell>150</cell><cell>207</cell><cell>228</cell></row><row><cell>UK Universities</cell><cell>Q1094046</cell><cell>64</cell><cell>108</cell><cell>208</cell><cell>224</cell></row><row><cell>Gene Wiki</cell><cell>Q30555</cell><cell>12</cell><cell>22</cell><cell>30</cell><cell>37</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5 .</head><label>5</label><figDesc>Numbers of predicates in the 2015 dump and WDumper could not fetch, by entity.</figDesc><table><row><cell cols="4">Entity schema:name skos:prefLabel Total</cell></row><row><cell>Q23</cell><cell>184</cell><cell cols="2">184 368</cell></row><row><cell>Q355643</cell><cell>23</cell><cell>23</cell><cell>46</cell></row><row><cell>Q1094046</cell><cell>22</cell><cell>22</cell><cell>44</cell></row><row><cell>Q30555</cell><cell>5</cell><cell>5</cell><cell>10</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6 .</head><label>6</label><figDesc>Number of qualifiers and references for the selected property of the selected entity in the output and input of WDumper (2020 dump).</figDesc><table><row><cell cols="2">Entity Property</cell><cell cols="4">Qualifiers Output Input Output Input References</cell></row><row><cell>Q23</cell><cell>P26</cell><cell>4</cell><cell>4</cell><cell>2</cell><cell>2</cell></row><row><cell>Q355643</cell><cell>P485</cell><cell>1</cell><cell>1</cell><cell>1</cell><cell>1</cell></row><row><cell>Q1094046</cell><cell>P355</cell><cell>1</cell><cell>1</cell><cell>1</cell><cell>1</cell></row><row><cell>Q17487737</cell><cell>P680</cell><cell>24</cell><cell>24</cell><cell>96</cell><cell>96</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">Query to count all data items https://w.wiki/yVY accessed 9 February 2021. Note that the query execution timesout if you try to return the count query SELECT (COUNT(DISTINCT ?0) AS ?numTopics) WHERE { ?s wdt:P31 ?o }.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">https://query.wikidata.org/sparql accessed February 2021</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">https://www.wikidata.org/wiki/Wikidata:Database_download</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_3">See this post: https://addshore.com/2019/10/your-own-wikidata-queryservice-with-no-limits/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_4">https://www.wikidata.org/wiki/Wikidata:Database_download#Partial_RDF_ dumps -accessed</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_5"> February 2021   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6">Note that in Wikidata, each statement can have a rank that can be used to identify the preferred value if the statement has more than one value. Ranks are available in Wikidata RDF data model like qualifiers and references.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_7">Downloaded from https://archive.org/download/wikidata-json-20150427 -ac-</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_8">cessed 11 November 2020 11 Downloaded from https://dumps.wikimedia.org/wikidatawiki/entities/ -accessed 15 November 2020</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_9">https://wikidata-todo.toolforge.org/stats.php</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_10">Downloaded from https://archive.org/download/wikidata-json-20150427 -accessed 20 December 2020</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_11"><ref type="bibr" target="#b14">15</ref> https://www.wikidata.org/wiki/Wikidata:History_Query_Service</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgement. We would like to acknowledge the fruitful discussions with the participants of project 35 of the BioHackathon-Europe 2020; Dan Brickley, Jose Emilio Labra Gayo, Eric Prud'hommaux, Andra Waagmeester, Harold Solbrig, Ammar Ammar, and Guillermo Benjaminsen.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="https://github.com/bennofs/wdumper" />
		<title level="m">Wdumper -a tool to create custom wikidata rdf dumps</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<ptr target="https://www.wikidata.org/wiki/Wikidata:WikiProject_Schemas/Subsetting" />
		<title level="m">WikiProject Schemas/Subsetting -Wikidata</title>
				<imprint>
			<date type="published" when="2020-12-31">2020-12-31</date>
		</imprint>
	</monogr>
	<note>Wikidata</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">DBpedia: A Nucleus for a Web of Open Data</title>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kobilarov</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-540-76298-0_52</idno>
		<ptr target="https://doi.org/10.1007/978-3-540-76298-052" />
	</analytic>
	<monogr>
		<title level="m">ISWC</title>
				<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">fixed syntax errors</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A H</forename><surname>Beghaeiraveri</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.4534445</idno>
		<ptr target="https://doi.org/10.5281/zenodo.4534445" />
		<imprint>
			<date type="published" when="2021-02">Feb 2021</date>
			<biblScope unit="page" from="27" to="31" />
		</imprint>
	</monogr>
	<note>Wikidata dump</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Wikidata Subsets and Specification Files Created by WDumper</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A H</forename><surname>Beghaeiraveri</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.4495855</idno>
		<ptr target="https://doi.org/10.5281/zenodo.4495855" />
		<imprint>
			<date type="published" when="2021-02">Feb 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A platform for scalable, collaborative, structured information integration</title>
		<author>
			<persName><forename type="first">K</forename><surname>Bollacker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tufts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pierce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cook</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IIWeb&apos;07</title>
				<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E L</forename><surname>Gayo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ammar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Brickley</surname></persName>
		</author>
		<idno type="DOI">10.37044/osf.io/wu9et</idno>
		<ptr target="https://doi.org/10.37044/osf.io/wu9et" />
		<title level="m">Knowledge graphs and wikidata subsetting</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E L</forename><surname>Gayo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Prud'hommeaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Boneva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kontokostas</surname></persName>
		</author>
		<title level="m">Validating RDF data</title>
				<imprint>
			<publisher>Morgan &amp; Claypool Publishers</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">7</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Schema. org: evolution of structured data on the web</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">V</forename><surname>Guha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Brickley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Macbeth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">59</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="44" to="51" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Hogan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Blomqvist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cochez</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2003.02320[cs</idno>
		<ptr target="http://arxiv.org/abs/2003.02320" />
		<title level="m">Knowledge Graphs</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Khrouf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Troncy</surname></persName>
		</author>
		<title level="m">Eventmedia: A lod dataset of events illustrated with media</title>
				<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Geo linked data</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">J</forename><surname>Lopez-Pellicer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>DEXA</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Mapping rdf graphs to property graphs</title>
		<author>
			<persName><forename type="first">S</forename><surname>Matsumoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Yamanaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chiba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1812.01801</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Domain specific knowledge graph embedding for analogical link discovery</title>
		<author>
			<persName><forename type="first">N</forename><surname>Mimouni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Moissinac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tuan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Intelligent Systems</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Yago 4: A reason-able knowledge base</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">P</forename><surname>Tanon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Weikum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Suchanek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Semantic Web Conference</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="583" to="596" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">UniProt Consortium: Uniprot: a hub for protein information</title>
	</analytic>
	<monogr>
		<title level="j">NAR</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="issue">D1</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Wikidata: a free collaborative knowledgebase</title>
		<author>
			<persName><forename type="first">D</forename><surname>Vrandečić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krötzsch</surname></persName>
		</author>
		<idno type="DOI">10.1145/2629489</idno>
		<ptr target="https://doi.org/10.1145/2629489" />
	</analytic>
	<monogr>
		<title level="j">CACM</title>
		<imprint>
			<biblScope unit="volume">57</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page" from="78" to="85" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Wikidata as a knowledge graph for the life sciences</title>
		<author>
			<persName><forename type="first">A</forename><surname>Waagmeester</surname></persName>
		</author>
		<idno type="DOI">10.7554/eLife.52614</idno>
		<ptr target="https://doi.org/10.7554/eLife.52614" />
	</analytic>
	<monogr>
		<title level="j">eLife</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
