<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">SectionLinks: Mapping Orphan Wikidata Entities onto Wikipedia Sections</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Natalia</forename><surname>Ostapuk</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Fribourg</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Djellel</forename><surname>Difallah</surname></persName>
							<email>djellel@nyu.edu</email>
							<affiliation key="aff1">
								<orgName type="institution">New York University</orgName>
								<address>
									<region>New York</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Philippe</forename><surname>Cudré-Mauroux</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Fribourg</orgName>
								<address>
									<settlement>Fribourg</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">SectionLinks: Mapping Orphan Wikidata Entities onto Wikipedia Sections</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">1E98EEF1DB5212671AB205E58BF1BD40</idno>
					<idno type="DOI">10.5281/zenodo.3840622</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T06:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Wikidata</term>
					<term>Wikipedia</term>
					<term>Linked Data</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Wikidata is a key resource for the provisioning of structured data on several Wikimedia projects, including Wikipedia. By design, all Wikipedia articles are linked to Wikidata entities; such mappings represent a substantial source of both semantic and structural information. However, only a small subgraph of Wikidata is mapped in that way --only about 10% of the sitelinks are linked to English Wikipedia, for example. In this paper, we describe a resource we have built and published to extend this subgraph and add more links between Wikidata and Wikipedia. We start from the assumption that a number of Wikidata entities can be mapped onto Wikipedia sections, in addition to Wikipedia articles. The resource we put forward contains tens of thousands of such mappings, hence considerably enriching the highly structured Wikidata graph with encyclopedic knowledge from Wikipedia.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Knowledge Graphs (KGs) provide a rich, structured, and multilingual source of information useful for a variety of applications that require machinereadable data. KGs are leveraged in search engines, natural language understanding, and virtual assistants, to name but a few examples. A KG is usually represented as a graph of vertices denoting entities and connected with directed edges depicting their relationships. KGs can be constructed automatically using information extraction techniques, or semi-automatically, as is the case with Wikidata 3 , a KG built and maintained by a community of volunteers. Wikidata has the advantage of being curated by humans and of being tightly integrated with multiple Wikimedia projects (e.g., Wikipedia, Wikimedia Commons, and Wiktionary). For example, every Wikipedia article across all languages has a corresponding and unique language-independant Wikidata entity. This mapping between Wikipedia and Wikidata is beneficial for both projects. On one hand, it facilitates information extraction and standardization of Wikipedia articles across languages, which can benefit from the standard structure and values of their Wikidata counterpart, e.g., for populating infoboxes. On the other hand, Wikipedia articles are routinely updated, which in turn keeps Wikidata fresh and useful for online applications.</p><p>However, the Wikipedia editorial guidelines require that an entity be notable or worthy of notice to be added to the encyclopedia, which is not the case of Wikidata. Hence, only a fraction of Wikidata entities has a corresponding article in any language. We refer to the remaining entities, without an article, as orphans. In the absence of a textual counterpart, orphans often suffer from incompleteness and lack of maintenance.</p><p>Our present work stems from the observation that a substantial number of orphan entities are indeed available in Wikipedia, but not at the page level; orphan entities can be described within existing Wikipedia articles in the form of sections, subsections, and paragraphs of a more generic concept or fact. Interestingly, even a short section describing an orphan Wikidata entity can carry useful information that could enrich the entity with additional facts and relationships. Such pieces of information are unfortunately buried inside long articles without direct relevance to the main subject. Instead, we propose to establish a fine-grained mapping between Wikidata orphan entities and Wikipedia (sub)-sections.</p><p>Our main contribution is a dataset of such mapping between Wikidata and Wikipedia sections that we created using several algorithmic methods, ranging from string matching to graph inference.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>To the best of our knowledge, we are the first to come up with a resource providing fine-grained mappings between Wikipedia and Wikidata; our mappings come in addition to the existing links that Wikipedia provides to Wikidata through section anchors (see Section 3).</p><p>A similar effort of matching entities to Wikipedia articles was made by Tonon et al. in <ref type="bibr" target="#b17">[18]</ref>. The paper addresses the problem of constructing a knowledge graph of Web Entities and mapping it onto DBpeadia with Wikipedia articles acting as DBpedia entries.</p><p>Our effort is not directly related to link prediction <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b11">12]</ref>, which typically operates in a homogeneous domain (e.g., when trying to infer new links in a given social network or knowledge graph), while we operate across two heterogeneous domains (i.e., Wikidata and Wikipedia). It is however related to Ad-hoc Object Retrieval techniques <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b16">17]</ref>, which retrieve target entities based on keyword or natural language queries, as well as to Entity Linking <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b0">1,</ref><ref type="bibr">11]</ref>, which attempts to link mentions in Web text to their referent entities in a knowledge base.</p><p>A special case of Entity Linking is Wikipedia Linking, which aims at discovering links between Wikipedia documents. This task was broadly studied within the Wiki track of INEX<ref type="foot" target="#foot_0">4</ref> conference <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>. Participants were invited to establish links between Wikipedia articles both at the page and text level (i.e. detect an anchor point in the text of the source document and a best entry point in the text of the target). The task of linking documents at the text level is of particular interest to us as it is a general case of linking a document to a section and closely relates to the main topic of this paper. A number of interesting approaches were developed both for identifying link source and target pages <ref type="bibr" target="#b7">[8]</ref> and detecting the best entry point inside the text of the target <ref type="bibr" target="#b2">[3]</ref>.</p><p>Our work is also directly related to information extraction <ref type="bibr" target="#b14">[15]</ref> and KG construction <ref type="bibr" target="#b13">[14]</ref> efforts. In that context, a number of systems have recently been proposed to extract information, often in the form of triples, from structured or unstructured content and link it to a semi-structured representation like a knowledge graph. DeepDive <ref type="bibr" target="#b18">[19]</ref>, for instance, is a well-known tool that employs statistical learning, inference and declarative constructs written by the user to build a knowledge base from a large collection of text. FRED <ref type="bibr" target="#b1">[2]</ref> is a machine-reading tool that automatically generates RDF/OWL ontologies and linked data from multilingual natural language text. MapSDI <ref type="bibr" target="#b8">[9]</ref> is a recent rule-based mapping framework for integrating heterogeneous data into knowledge graphs. It takes as input a set of data sources and mapping rules to produce a consolidated knowledge graph. None of those tools is readily applicable to our problem of linking Wikipedia sections to Wikidata entities, however.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Relevance and Use Cases</head><p>In Wikidata, entities are characterized by a unique identifier (a sequential integer prefixed with Q), multilingual labels, descriptions, and aliases when available. Each entity may have multiple statements to express a property or a relationship with another entity. An entity can have Sitelinks <ref type="foot" target="#foot_1">5</ref> referencing other Wikimedia projects. These are hyperlinks that establish an identity mapping between the entity and, for instance, a Wikipedia page. Thanks to Sitelinks, Wikidata is often utilized as a hub for multilingual data, connecting a given concept to articles written in a dozen languages.</p><p>To understand Wikidata's Sitelinks coverage, we collected the number of label entities per language. We focus on the 15 languages having +1 Million Wikipedia articles (see Section 4). We examined the number of orphan entities (defined above in Section 1) having a label in each language, as shown in Figure <ref type="figure" target="#fig_0">1</ref>, which we contrast to the number of available Wikipedia articles. We see that the gap between the number of orphans and articles is much higher for languages having more labels. In fact, English Wikipedia, the largest and most active project of all Wikis, links to only about 10% of all Wikidata entities having an English label. This discrepancy signals a necessity to close this gap using alternate methods.</p><p>This work aims to identify potential orphan entity textual content that may exist within Wikipedia in the form of sections. This content can be linked using anchor links to article sections. Currently, Wikidata does not support using anchors links as a Sitelink, i.e. linking to a specific section of a page. It is worth noting that inter-language Wikipedia can perform this operation, for example, the Wikidata entity Q2915096 contains a Sitelink to the English Wikipedia page Survival_function, and all the other Wikidata sitelinks are listed on this page in the left column (Figure <ref type="figure">2</ref>). A link to a section can be added to this list and thus can be mapped to the source Wikidata entity, as is the case for the French language. Unfortunately, this is done inconsistently and provides only an indirect mapping to Wikidata, and also assumes that at least one language has a dedicated Wikipedia page for the entity. Our proposed resource fills this important gap by building an external resource to map Wikidata orphans to Wikipedia, without entering a sitelink.</p><p>Figure <ref type="figure">3</ref> illustrates what we want to achieve. It depicts the Wikipedia entry for Brexit. While the Brexit entity (Q7888194) from Wikidata correctly links to that page, two related Wikidata entities are orphans: European Union (Withdrawal) Act 2018 (Q29582790) and exit day (Q59189602). Linking those two entities to their corresponding sections in Wikipedia, as shown in the figure, would provide important information and context to Wikidata and greatly improve a number of key downstream tasks such as ad-hoc object retrieval, joint embeddings, or question answering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">The Dataset</head><p>We developed two different algorithms to derive mappings from Wikidata entities to Wikipedia sections. We ran both our algorithms on 15 languages and obtained tens of thousands of new links in the process (see section 4.2 for details). The two resulting datasets complement each other (i.e. contain sitelinks for different sets of entities) and are both available as part of our resource. The rest of this section describes our methods and results in detail, and provides performance numbers and illustrative examples to better assess the usefulness of our resource.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data Generation Pipeline</head><p>We consider a bipartite graph G whose vertices consist of two disjoint subsets: D, representing Wikidata entities that are missing a Wikipedia link, and P, representing Wikipedia page sections. Our goal is to correctly match as many vertices as possible from D to P (i.e., to create as many correct links as possible between Wikidata entities and Wikipedia section). To help with this task, we use existing labels and statements available from each entity, as well as the section titles that we collect from the 15 Wikis. We proceed in four steps:</p><p>Candidates selection: the first step is to identify candidates, both from Wikidata (D vertices) and Wikipedia (P vertices), in order to create the matching graph G.</p><p>Key generation: then, we create a key (or a set of keys) to represent each vertex in D and P.</p><p>Matching: at this stage, we create candidate links between by matching keys in D with keys in P.</p><p>Filtering: finally, as the matching step may result in many false positive links, we consider a postprocesing step where each resulting link is vetted against a set of roles or conditions.</p><p>We describe two different instantiations of our data generation pipeline below: one considering a strict all-to-all matching algorithm, and a second, graph-based algorithm that takes into account the neighborhood of each candidate.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>All-to-All Matching Algorithm</head><p>Our first approach considers a complete bipartite graph, where each Wikidata entity in D is matched to all Wikipedia sections in P. Since we do not apply any restriction on the candidate targets (Wikipedia pages) and since the number of matches grows quadratically, the key comparison method and the filtering functions both have to be very strict, otherwise the algorithm would return a lot of false positive matches. We achieve this with the requirement that a Wikipedia key has to comprise all tokens from both the page title and the section title; as such, a Wikipedia key is specific enough to guarantee with a high probability that it refers to the same object as a corresponding Wikidata entity.</p><p>Candidates selection First, we identify all orphan Wikidata entities, i.e. all entities that have a label in a given language but do not have a sitelink to a corresponding Wikipedia page or section. Orphans are further filtered by type to exclude service pages like categories or templates, as well as some types which have homonymous labels but rarely match any Wikipedia section (for example, an entity of the type painting with the label The Crucifixion match the Wikipedia section describing the crucifixion of Jesus, which is irrelevant to this object).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Key generation We consider a set of keys for each Wikidata candidate in D.</head><p>This set of keys consists of its label and all its aliases for a given language. For example, for an entity Q63854053 the set of keys will be {"spun silk", "noil", "silk noil"}. To generate keys for Wikipedia page section in P, we concatenate the page title with the section title. After all keys are generated, we split each key into tokens, remove punctuation and stop-words, sort tokens in alphabetic order and concatenate them back together. We used stop word lists provided by the NLTK package<ref type="foot" target="#foot_2">6</ref> in this context.</p><p>Matching The output of the key generation step consist of two key-value tables: one for Wikidata entities, where the keys are as described above and each value is an entity id, and another for Wikipedia sections, with (page_title−section_title) pairs as values. These two tables are then joined by key and grouped by QID (Wikidata entity id). This operation was performed on a Hadoop cluster.</p><p>Filtering The last step of the pipeline is result filtering. As mentioned above, this approach considers all possible matches and hence may bring up a lot of false positives, therefore the filtering function we use is strict also: we keep only those QIDs for which exactly one Wikipedia section was found. In more formal terms, this step of the algorithm checks the output of the groupBy operation and filters out records which grouped more than one value per QID. Figure <ref type="figure">4</ref> outlines the overall pipeline of our first approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 4. All-to-all matching pipeline</head><p>Neighbors Matching Algorithm Although the above algorithm demonstrates a good performance (over 80% precision for English Wikipedia), a manual analysis reveals that it has a relatively low recall. The reason is that in many cases, when a Wikipedia section describes an object its title is selfsufficient, i.e. it stands on its own and does not depend on the title page to identify the object. Hence, matching a Wikidata label strictly with a combination of a page and section title results in a lot of false negatives as the page title introduces redundancy. On the other hands, matching with section titles only will significantly drop the precision in general. To tackle this problem, we restrict the set of candidate Wikipedia sections for each Wikidata entity by leveraging the Wikidata graph structure.</p><p>Candidates selection Our candidate selection algorithm in this case is based on the assumption that a Wikipedia page that is "semantically" related (e.g., to a subclass relation) to a Wikidata entity in D is more likely to contain a section relevant to that entity.</p><p>We introduce a second condition to further restrict the candidates as follows: a candidate in P should be related to one and only one source entity in D for a particular edge type. For example, consider an orphan entity badminton racket and a triple [(badminton), (uses), (badminton racket)]. Here, (badminton) is a good candidate, because it is linked to (badminton racket) with the relation [uses]. On the other hands, in the triple [(Sofia Shinas), (occupation), (singer)], (singer) is not an interesting candidate for (Sofia Shinas), as many entities have an occupation singer.</p><p>As such, we developed the following algorithm pipeline for selecting candidate Wikipedia sections using a graph-based approach:</p><p>-Identify an orphan Wikidata entity; -Collect its neighbor entities following all incoming and outgoing edges; -Filter out neighbors that do not have a Wikipedia sitelink; -Filter out neighbors with non-unique edges; -Extract Wikipedia sitelinks from the remaining neighbors; -Consider sections of the resulting Wikipedia pages as candidates for matching.</p><p>This algorithms yields excellent results in practice as we describe below in Section 4.3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Key generation</head><p>As we significantly limited the set of candidate Wikipedia sections in P, we consider a different way of constructing the keys. First, we do not always consider the tokens from the page title for the keys in P (although they may be included). Second, in addition to removing stop words and punctuation, we consider a third postprocessing step by stemming the key tokens, i.e. we remove all affixes that mostly carry morphological information and only keep just their root (e.g. words works, worked, working are all reduced to work ). Finally, we remove disambiguation tokens from Wikipedia page titles: when a title is ambiguous, a disambiguation word or phrase can be added in parenthesis. For example, titles Mercury (element), Mercury (planet) and Mercury (mythology) are all reduced to Mercury.</p><p>Matching The matching step is similar to the one in our first algorithm, but instead of running a join of two tables we process each Wikidata entity in D individually. If one of the Wikidata keys exactly match a Wikipedia section key, we consider the section as a potential sitelink for this entity.</p><p>Filtering Due to the various manipulations we consider on the keys, we may end up with situations where different Wikipedia sections have the same key that matches a Wikidata key. For a example, an article Rotterdam Metro includes two sections: Line D and Lines. After stemming and stop words removal, both section titles are reduced to line. If we consider a Wikidata entity with a label Line D (which is also reduced to line), we get two potential matches. In that case, we consider the edit distance between the Wikidata label and the section title (in their original forms) and pick the closest match to break the tie. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Resource Description</head><p>We ran our methods using the dumps of 15 Wikipedias obtained from April 2020. <ref type="foot" target="#foot_3">7</ref> While the Wikidata graph dump was obtained from February 2020. <ref type="foot" target="#foot_4">8</ref>The final resource contains 126,151 sitelinks for 109,734 unique entities for 15 languages obtained with the two methods described above. The subset of languages we initially considered were chosen according to the following criteria:</p><p>1. Number of articles in the corresponding Wikipedia (over 1 million) <ref type="foot" target="#foot_5">9</ref>2. Number of Wikipedia active users (over 1000)</p><p>We plan to run algorithms on more languages in the future. We report below the full list of languages we considered as well as detailed statistics on the datasets (see Table <ref type="table" target="#tab_0">1</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Evaluation Results</head><p>To estimate the precision of our resource, we randomly sampled several hundreds matches from each dataset and manually evaluated them as either true or false. For instance, one algorithm matched the entity Q49001814 (Timber Dam, the name of the dam in Montana, USA) to the section Timber dams of the Wikipedia page Dam, which describes a type of dams made of timber. This match was labelled as a false positive. An example of a true positive match is a mapping of the entity Q334415 (security camera) onto the page Surveillance, section Cameras. We labelled each sample this way and then divided the number of true positive matches by the sample size to get a precision value. We then generalized from the sample observations to the whole dataset using linear extrapolation in order to estimate the dataset precision. Table <ref type="table" target="#tab_1">2</ref> reports our results. We evaluated 12 samples -one sample per algorithm plus the joint results, for 4 different languages (Arabic, English, French, Russian). Each sample contains around 200 mappings. This number was chosen empirically, as we observed that 200 random examples were enough to stabilize the metric, and increasing the sample size did not change the resulting value significantly. Overall we manually labelled 2400 mappings.</p><p>Our evaluation aims to demonstrate that the overall accuracy of the resource is high enough that it can be used for many tasks that do not require a perfect dataset (for example, most deep learning algorithms are robust to errors in the training set). We cannot provide evaluation results for all languages unfortunately, as we decided to focus on those languages we felt comfortable with only.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Availability and Reusability</head><p>Our resource is available in JSON and RDF formats and comply with the Wikibase data model. <ref type="foot" target="#foot_6">10</ref> To keep the resource compact and as easy to process as possible, we only publish the sitelinks discover using our methods.</p><p>In the JSON representation, an entity contains two fields: id (the unique identifier of an entity) and sitelinks (links to Wikipedia pages). Each sitelink record comprises three fields: site, title and url. A section title is appended to the page title separated with # symbol. Such a compound title is then URL-encoded and added to the URL path. Following the Wikidata guidelines, each entity is encoded as a single line.</p><p>The RDF dump is serialized using the Turtle format and stores nodes describing Wikipedia links. Section titles are added in the same manner as described above . <ref type="foot" target="#foot_7">11</ref>The resource is published on the Zenodo platform under CC BY 4.0 license. <ref type="foot" target="#foot_8">12</ref> . The canonical citation is available on the Zenodo page. The source code is also available on our github repository to help maintain and generate newer releases in the future. 13 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion and Future Work</head><p>We presented a dataset that extends Wikidata orphan entities with Sitelinks referencing Wikipedia sections for the 15 most prominent languages in Wikipedia. To generate this resource, we employed string matching and graph processing methods that leverage multilingual labels and the graph structure to find corresponding sections in Wikipedia. Since our methods use heuristics, we compute the accuracy of a subset of the data using manual judgment. This piece of information can be useful to inform downstream application on how to use the data. For instance, for entities with an English label, we identified 9,834 links with 82% accuracy when using exact label matching, and 25,469 links when using graph-based method alone with 81% accuracy.</p><p>We believe that using this resource can improve both sources in terms of completeness and freshness, as well as diminish the information gap that persists between Wikipedia-based entities and tail-entities. For example, one could build targeted information extraction tools and automatically curate entities that do not have a dedicated Wikipedia article using our resource. As future work, we plan to incorporate embedding-based similarity scores into our mapping method and perform a comprehensive evaluation of the obtained results in terms of both precision and recall. We also envision building a section recommendation system that can be offered to Wikidata editors for relevance judgment.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Language statistics and gaps in Wikidata.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .Fig. 3 .</head><label>23</label><figDesc>Fig. 2. A list of Wikidata sitelinks on a Wikipedia page. The French link points to: https://fr.wikipedia.org/wiki/Analys _ de _ survie#Fonction _ de _ survie</figDesc><graphic coords="4,169.35,489.79,276.67,138.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="8,169.35,115.84,276.66,196.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Datasets size per language</figDesc><table><row><cell></cell><cell>All-to-all</cell><cell>Neighbors</cell><cell>Final</cell></row><row><cell></cell><cell>matching</cell><cell>matching</cell><cell></cell></row><row><cell>Arabic</cell><cell>672</cell><cell>1752</cell><cell>1792</cell></row><row><cell>Chinese</cell><cell>38</cell><cell>848</cell><cell>882</cell></row><row><cell>Dutch</cell><cell>2699</cell><cell>5230</cell><cell>5644</cell></row><row><cell>English</cell><cell>9834</cell><cell>25469</cell><cell>30351</cell></row><row><cell>French</cell><cell>6573</cell><cell>10436</cell><cell>13098</cell></row><row><cell>German</cell><cell>3923</cell><cell>9140</cell><cell>10443</cell></row><row><cell>Italian</cell><cell>8512</cell><cell>27775</cell><cell>32262</cell></row><row><cell>Japanese</cell><cell>215</cell><cell>4105</cell><cell>4229</cell></row><row><cell>Polish</cell><cell>663</cell><cell>678</cell><cell>1271</cell></row><row><cell>Portuguese</cell><cell>399</cell><cell>1188</cell><cell>1407</cell></row><row><cell>Russian</cell><cell>2844</cell><cell>3404</cell><cell>4675</cell></row><row><cell>Spanish</cell><cell>3081</cell><cell>6754</cell><cell>7939</cell></row><row><cell>Swedish</cell><cell>2552</cell><cell>3035</cell><cell>3343</cell></row><row><cell>Ukrainian</cell><cell>309</cell><cell>290</cell><cell>571</cell></row><row><cell>Vietnamese</cell><cell>8164</cell><cell>108</cell><cell>8244</cell></row><row><cell>Total</cell><cell>50478</cell><cell>100212</cell><cell>126151</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Datasets precision</figDesc><table><row><cell></cell><cell>All-to-all</cell><cell>Neighbors</cell><cell>Final</cell></row><row><cell></cell><cell>matching</cell><cell>matching</cell><cell></cell></row><row><cell>Arabic</cell><cell>0.99</cell><cell>1.0</cell><cell>0.99</cell></row><row><cell>English</cell><cell>0.82</cell><cell>0.81</cell><cell>0.82</cell></row><row><cell>French</cell><cell>0.85</cell><cell>0.92</cell><cell>0.89</cell></row><row><cell>Russian</cell><cell>0.88</cell><cell>0.85</cell><cell>0.87</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">International Workshop of the Initiative for the Evaluation of XML Retrieval</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">https://www.wikidata.org/wiki/Help:Sitelinks</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">https://www.nltk.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_3">https://dumps.wikimedia.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_4">https://dumps.wikimedia.org/wikidatawiki/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_5">https://meta.wikimedia.org/wiki/List _ of _ Wikipedias</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_6">https://www.mediawiki.org/wiki/Wikibase/DataModel</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_7">For the detailed description of the Wikidata RDF format refer to: https://www. mediawiki.org/wiki/Wikibase/Indexing/RDF _ Dump _ Format</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_8">http://doi.org/10.5281/zenodo.3840622</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This project has received funding from the European Research Council (ERC) under the European Union's Horizon 2020 research and innovation programme (grant agreement 683253/GraphInt).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Zencrowd: Leveraging probabilistic reasoning and crowdsourcing techniques for large-scale entity linking</title>
		<author>
			<persName><forename type="first">G</forename><surname>Demartini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Difallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cudré-Mauroux</surname></persName>
		</author>
		<idno type="DOI">10.1145/2187836.2187900</idno>
		<idno>2187836.2187900</idno>
		<ptr target="https://doi.org/10.1145/2187836.2187900" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st International Conference on World Wide Web</title>
				<meeting>the 21st International Conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="469" to="478" />
		</imprint>
	</monogr>
	<note>WWW &apos;12</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Semantic web machine reading with FRED</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gangemi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Presutti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Recupero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">G</forename><surname>Nuzzolese</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Draicchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mongiovì</surname></persName>
		</author>
		<idno type="DOI">10.3233/SW-160240</idno>
		<ptr target="https://doi.org/10.3233/SW-160240" />
	</analytic>
	<monogr>
		<title level="j">Semantic Web</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="873" to="893" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Link discovery in the wikipedia</title>
		<author>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">X</forename><surname>Tang</surname></persName>
		</author>
		<editor>Shlomo Geva, Jaap Kamps</editor>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>Andrew Trotman</publisher>
			<biblScope unit="page">326</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Collective entity linking in web text: A graphbased method</title>
		<author>
			<persName><forename type="first">X</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.1145/2009916.2010019</idno>
		<ptr target="https://doi.org/10.1145/2009916.2010019" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="765" to="774" />
		</imprint>
	</monogr>
	<note>SIGIR &apos;11</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Overview of the INEX 2008 link the wiki track</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W C</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-03761-0_32</idno>
		<idno>-642-03761-0_32</idno>
		<ptr target="https://doi.org/10.1007/978-3-642-03761-0_32" />
	</analytic>
	<monogr>
		<title level="m">Advances in Focused Retrieval, 7th International Workshop of the Initiative for the Evaluation of XML Retrieval, INEX 2008</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</editor>
		<meeting><address><addrLine>Dagstuhl Castle, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">December 15-18, 2008. 2008</date>
			<biblScope unit="volume">5631</biblScope>
			<biblScope unit="page" from="314" to="325" />
		</imprint>
	</monogr>
	<note>Revised and Selected Papers</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Overview of the INEX 2009 link the wiki track</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W C</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-14556-8_31</idno>
		<idno>-642-14556-8_31</idno>
		<ptr target="https://doi.org/10.1007/978-3-642-14556-8_31" />
	</analytic>
	<monogr>
		<title level="m">Focused Retrieval and Evaluation, 8th International Workshop of the Initiative for the Evaluation of XML Retrieval, INEX 2009</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</editor>
		<meeting><address><addrLine>Brisbane, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2009">December 7-9, 2009. 2009</date>
			<biblScope unit="volume">6203</biblScope>
			<biblScope unit="page" from="312" to="323" />
		</imprint>
	</monogr>
	<note>Revised and Selected Papers</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Overview of INEX 2007 link the wiki track</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W C</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-540-85902-4_32</idno>
		<idno>-540-85902-4_32</idno>
		<ptr target="https://doi.org/10.1007/978-3-540-85902-4_32" />
	</analytic>
	<monogr>
		<title level="m">Focused Access to XML Documents, 6th International Workshop of the Initiative for the Evaluation of XML Retrieval, INEX 2007</title>
		<title level="s">Selected Papers. Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">N</forename><surname>Fuhr</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Lalmas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</editor>
		<meeting><address><addrLine>Dagstuhl Castle, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">December 17-19, 2007. 2007</date>
			<biblScope unit="volume">4862</biblScope>
			<biblScope unit="page" from="373" to="387" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">University of waterloo at INEX2007: adhoc and link-the-wiki tracks</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Y</forename><surname>Itakura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L A</forename><surname>Clarke</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-540-85902-4_35</idno>
		<idno>- 540-85902-4_35</idno>
		<ptr target="https://doi.org/10.1007/978-3-540-85902-4_35" />
	</analytic>
	<monogr>
		<title level="m">Focused Access to XML Documents, 6th International Workshop of the Initiative for the Evaluation of XML Retrieval, INEX 2007</title>
		<title level="s">Selected Papers. Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">N</forename><surname>Fuhr</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Lalmas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</editor>
		<meeting><address><addrLine>Dagstuhl Castle, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">December 17-19, 2007. 2007</date>
			<biblScope unit="volume">4862</biblScope>
			<biblScope unit="page" from="417" to="425" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Mapsdi: A scaled-up semantic data integration framework for knowledge graph creation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Jozashoori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vidal</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-33246-4_4</idno>
		<idno>030-33246-4_4</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-33246-4_4" />
	</analytic>
	<monogr>
		<title level="m">On the Move to Meaningful Internet Systems: OTM 2019 Conferences -Confederated International Conferences: CoopIS, ODBASE, C&amp;TC 2019</title>
		<title level="s">Proceedings. Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">H</forename><surname>Panetto</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Debruyne</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Hepp</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Lewis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Ardagna</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Meersman</surname></persName>
		</editor>
		<meeting><address><addrLine>Rhodes, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">October 21-25, 2019. 2019</date>
			<biblScope unit="volume">11877</biblScope>
			<biblScope unit="page" from="58" to="75" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The link prediction problem for social networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liben-Nowell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Kleinberg</surname></persName>
		</author>
		<idno type="DOI">10.1145/956863.956972</idno>
		<ptr target="https://doi.org/10.1145/956863" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2003 ACM CIKM International Conference on Information and Knowledge Management</title>
				<meeting>the 2003 ACM CIKM International Conference on Information and Knowledge Management<address><addrLine>New Orleans, Louisiana, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2003">November 2-8, 2003. 2003</date>
			<biblScope unit="page" from="556" to="559" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Entity linking at web scale</title>
		<author>
			<persName><forename type="first">T</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><surname>Mausam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Etzioni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Joint Workshop on Automatic Knowledge Base Construction and Web-Scale Knowledge Extraction</title>
				<meeting>the Joint Workshop on Automatic Knowledge Base Construction and Web-Scale Knowledge Extraction<address><addrLine>USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="84" to="88" />
		</imprint>
	</monogr>
	<note>AKBC-WEKEX &apos;12</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A survey of link prediction in complex networks</title>
		<author>
			<persName><forename type="first">V</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Berzal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C C</forename><surname>Talavera</surname></persName>
		</author>
		<idno type="DOI">10.1145/3012704</idno>
		<ptr target="https://doi.org/10.1145/3012704" />
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">33</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Ad-hoc object retrieval in the web of data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pound</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zaragoza</surname></persName>
		</author>
		<idno type="DOI">10.1145/1772690.1772769</idno>
		<idno>1772690.1772769</idno>
		<ptr target="http://doi.acm.org/10.1145/1772690.1772769" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 19th International Conference on World Wide Web</title>
				<meeting>the 19th International Conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="771" to="780" />
		</imprint>
	</monogr>
	<note>WWW &apos;10</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Knowledge graph construction techniques</title>
		<author>
			<persName><forename type="first">L</forename><surname>Qiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhiguang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Computer Research and Development</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="582" to="600" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Information extraction</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sarawagi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and trends in databases</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="261" to="377" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Entity linking with a knowledge base: Issues, techniques, and solutions</title>
		<author>
			<persName><forename type="first">W</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Han</surname></persName>
		</author>
		<idno type="DOI">10.1109/TKDE.2014.2327028</idno>
		<ptr target="https://doi.org/10.1109/TKDE.2014.2327028" />
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Knowl. Data Eng</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="443" to="460" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Combining inverted indices and structured search for ad-hoc object retrieval</title>
		<author>
			<persName><forename type="first">A</forename><surname>Tonon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Demartini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cudré-Mauroux</surname></persName>
		</author>
		<idno type="DOI">10.1145/2348283.2348304</idno>
		<idno>2348283.2348304</idno>
		<ptr target="http://doi.acm.org/10.1145/2348283.2348304" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 35th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 35th International ACM SIGIR Conference on Research and Development in Information Retrieval<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="125" to="134" />
		</imprint>
	</monogr>
	<note>SIGIR &apos;12</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Tonon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Felder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Difallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cudré-Mauroux</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-319-46547-0_23</idno>
		<ptr target="http://www.slideshare.net/eXascaleInfolab/voldemortkg-mapping-schemaorg-and-web-entities-to-linked-open-data" />
		<title level="m">Voldemor-tKG: Mapping schema.org and Web Entities to Linked Open Data</title>
				<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Deepdive: declarative knowledge base construction</title>
		<author>
			<persName><forename type="first">C</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ré</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Cafarella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1145/3060586</idno>
		<ptr target="https://doi.org/10.1145/3060586" />
	</analytic>
	<monogr>
		<title level="j">Commun. ACM</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="93" to="102" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
