<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Topic-Sensitive Model for Salient Entity Linking</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Lei</forename><surname>Zhang</surname></persName>
							<email>l.zhang@kit.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Institute AIFB</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology (KIT)</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cong</forename><surname>Liu</surname></persName>
							<email>cong.liu@student.kit.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Institute AIFB</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology (KIT)</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Achim</forename><surname>Rettinger</surname></persName>
							<email>rettinger@kit.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Institute AIFB</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology (KIT)</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Topic-Sensitive Model for Salient Entity Linking</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B00A938CC68A68F12DB186B6DC7E79C2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T09:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In recent years, the amount of entities in large knowledge bases available on the Web has been increasing rapidly. Such entities can be used to bridge textual data with knowledge bases and thus help with many tasks, such as text understanding, word sense disambiguation and information retrieval. The key issue is to link the entity mentions in documents with the corresponding entities in knowledge bases, referred to as entity linking. In addition, for many entity-centric applications, entity salience for a document has become a very important factor. This raises an impending need to identify a set of salient entities that are central to the input document. In this paper, we introduce a new task of salient entity linking and propose a graph-based disambiguation solution, which integrates several features, especially a topic-sensitive model based on Wikipedia categories. Experimental results show that our method significantly outperforms the state-of-the-art entity linking methods in terms of precision, recall and F-measure.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In recent years, large repositories of structured knowledge publicly available on the Web, such as Wikipedia, DBpedia, Freebase and YAGO, have become valuable resources for information extraction. In this regard, entity linking, which leverages such knowledge bases to link words or phrases in natural language text with the corresponding entities, has emerged as a topic of major interest.</p><p>The challenges of entity linking lie in entity recognition and disambiguation. The first stage serves to detect words or phrases in text, also called mentions, that are likely to denote entities; the second stage performs the disambiguation of the recognized mentions into entities. Many methods <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8]</ref> have been proposed to address the problems of entity disambiguation and linking. However, these methods do not take into account the actual importance of entities w.r.t. the topics of the input document. In this work, the relation between the candidate entities and their associated categories are utilized to opt the entities that are related to the document topics.</p><p>In addition, there is an impending need to identify a set of salient entities in a document that play an important role in the content of the document, which would help to better understand its meaning or aboutness <ref type="bibr" target="#b8">[9]</ref>. This paper focuses on the task of salient entity linking, especially the disambiguation of The rest of the paper is organized as follows. We start with an overview of our framework for salient entity linking in Sec. 2. The details of features and measures used for salient entity disambiguation are provided in Sec. 3. Based on them, we discuss the graph-based disambiguation utilizing a topic-sensitive model in Sec. 4. Evaluation results are then presented in Sec. 5, followed by the conclusions in Sec. 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Framework</head><p>Before we discuss our salient entity linking framework, we first formulate the task of entity linking and then introduce the problem of salient entity linking, an extension of the general entity linking task.</p><p>Definition 1 (Entity Linking). Let M = {m 1 , m 2 , . . . , m p } denote a set of entity mentions in a document D. Given a knowledge base KB containing a set of entities E = {e 1 , e 2 , . . . , e n }, the objective of entity linking is to determine the referent entities in KB for the mentions in M , where two functions are to be found. For entity recognition, the mentions need to be extracted from D, where a recognition function er : D → 2 M will be computed. The resulting mentions (i.e., a subset µ ⊆ M ) are then mapped to entities in KB, where a disambiguation function ed : µ → E must be derived.</p><p>Definition 2 (Salient Entity Linking). Given a knowledge based KB and a document D, the recognition function of salient entity linking is same as general entity linking, i.e., er : D → 2 M . For the set of mentions µ ⊆ M yielded by the recognition function, the disambiguation function ed : µ → E ∪ {Non-Salient}, which maps the set of mentions µ to entities in the KB or to non-salient entities, must be derived, where non-salient entities are entities with no focus of attention in D, i.e., the document D is really not about such entities.</p><p>An illustration of our salient entity linking framework consisting of several components is given in Fig. <ref type="figure" target="#fig_0">1</ref>. In the following, we first introduce the components w.r.t. general entity linking and then discuss its extension with the components for salient entity linking by utilizing a topic-sensitive model.</p><p>For both general and salient entity linking, the input text is first processed by entity recognition, which detects the boundaries of mentions without knowing the actual referent entities or whether they are salient or non-salient entities. Then these mentions serve as the input of entity disambiguation, which is the focus of this work since we do not aim to compare the method's ability to recognize entity names in the input text.</p><p>Given a detected mention, its candidate referent entities are extracted from the knowledge base. For entity disambiguation regarding general entity linking, our framework combines different features including prior mention importance, mention-entity compatibility and entity-entity coherence. The feature of prior mention importance assigns the prior importance to each detected mention as weight and it will be used as the initial evidence for graph-based disambiguation. While the local feature of mention-entity compatibility captures the most likely entity behind the mention and the entity that best fits the context, the global feature of entity-entity coherence collectively captures the linked entities in a document that are related to each other. These features are then employed by graph-based disambiguation based on a personalized PageRank algorithm.</p><p>To aim for effective salient entity linking, we first perform text classification on the input text using a multi-class support vector machine (SVM) classifier based on Wikipedia categories<ref type="foot" target="#foot_0">1</ref> aligned with the training corpus. For each category, we compute the category probability of the input document that serves as the feature of document-specific category importance. In addition, we compute the strength of entity-category association based on the depth between each candidate entity and its categories. Such features are then incorporated into graph-based disambiguation using a topic-sensitive PageRank algorithm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Features and Measures</head><p>In this section, we discuss the features and measures needed for salient entity disambiguation, while the graph model and algorithm will be presented in Sec. <ref type="bibr" target="#b3">4</ref>.</p><p>Prior Mention Importance. We employ the Wikipedia link structures for determining the prior mention importance. As each Wikipedia article describes an entity, article titles, redirect pages and link anchors can be used to refer to the entity. Based on the above sources, we extract all surface forms of entities.</p><p>For each mention m with the name m.s as surface form of an entity, we define the probability P (m.s) that captures how likely m.s refers to an entity as</p><formula xml:id="formula_0">P (m.s) = count link (m.s) count link (m.s) + count text (m.s)<label>(1)</label></formula><p>where count link (m.s) is the number of articles that contain m.s as anchor text and count text (m.s) is the number of articles where m.s appears as raw text.</p><p>Mention-Entity Compatibility. For each mention m and its candidate referent entity e, we calculate the semantic similarity SS(m, e) representing the local mention-entity compatibility of m and e as follows</p><formula xml:id="formula_1">SS(m, e) = α • LP (m, e) + β • CS(m, e)<label>(2)</label></formula><p>where LP (m, e) is the link probability of e for m and CS(m, e) is the context similarity between m and e, α and β are tunable parameters with α + β = 1.</p><p>The link probability LP (m, e) can be calculated using the probability P (e|m.s) capturing how likely the mention name m.s refers to the entity e as follows LP (m, e) = P (e|m.s) = count link (e, m.s)</p><formula xml:id="formula_2">ei∈Em.s count link (e i , m.s)<label>(3)</label></formula><p>where count link (e, m.s) denotes the number of links using m. </p><p>Entity-Entity Coherence. The disambiguation is based on the feature of entity-entity coherence, which collectively captures the referent entities of the mentions contained in the same document that are related to each other. In this regard, we calculate the semantic relatedness between each pair of entities e i and e j by adopting the Wikipedia link-based measure described in <ref type="bibr" target="#b9">[10]</ref>, which is originally modeled after the Normalized Google Distance (NGD) <ref type="bibr" target="#b10">[11]</ref>, as follows</p><formula xml:id="formula_4">SR(e i , e j ) = 1 − log(max(|E i |, |E j |)) − log(|E i ∩ E j |) log(|E|) − log(min(|E i |, |E j |))<label>(5)</label></formula><p>where E i and E j are the sets of entities that link to e i and e j in KB respectively, and E is the set of all entities in KB. Document-specific Category Importance. For text classification of the input document, we employ John C. Platt's sequential minimal optimization for training a support vector machine (SVM) classifier <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>. Multi-category problems are solved using pairwise classification. To obtain proper probability estimates, we use the option that fits logistic regression models to the outputs of the SVM classifier. In our multi-category scenario, the predicted probabilities are coupled using Hastie and Tibshirani's pairwise coupling method <ref type="bibr" target="#b13">[14]</ref>. All these algorithms have been integrated into Weka 2 , a collection of machine learning algorithms for data mining tasks. Based on that, we calculate the category probability P (c i ) of the input text for each assigned category c i , which reflects the document-specific category importance.</p><p>Entity-Category Association. All candidate entities are mapped to the selected Wikipedia categories. In order to measure the entity-category association between an entity e and its assigned category c, we define the distance d(c, e) as the minimum depth at which the entity e is located in Wikipedia's category tree with the category c as the root. This is computed offline by performing a breadth-first search starting from the fundamental category that forms the root of Wikipedia's hierarchy to each entity. Then the semantic association SA(c, e) between entity e and category c can be calculated as</p><formula xml:id="formula_5">SA(c, e) = 1 d(c, e)<label>(6)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Graph Model and Algorithm</head><p>Based on the features and measures discussed in Sec. 3, we construct a directed weighted graph G = {N, R}, called disambiguation graph, where N = N M N E N C is the disjoint union of mention nodes N M , entity nodes N E and category nodes N C , and R is the set of directed edges representing relationships between these nodes. All detected mentions and their candidate referent entities are added into N M and N E , respectively, while the categories that the input text belongs to are added into N C . For each mention m and its candidate entity e, we add an edge from m to e into R. Additionally, we add an edge between e i and e j into R if they are connected in KB. Furthermore, for each association between an entity e and a category c, an edge from c to e will be added into R.</p><p>Once the disambiguation graph G is built, we apply a personalized PageRank algorithm <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref> over it. The calculation of the PageRank vector P r over G is equivalent to resolving the following equation</p><formula xml:id="formula_6">P r = d • T • P r + (1 − d) • v (<label>7</label></formula><formula xml:id="formula_7">)</formula><p>where T is the transition probability matrix, v is the initial evidence vector and d is the so called damping factor, usually set as 0.85. Each entry T ij in T is the evidence propagation ratio from node i to node j, which is computed in Eq. 8.</p><formula xml:id="formula_8">T ij =          SS(mi,ej ) k∈N E (i) SS(mi,e k ) if i ∈ N M , j ∈ N E SR(ei,ej ) k∈N E (i) SR(ei,e k ) if i ∈ N E , j ∈ N E SA(ci,ej ) k∈N E (i) SA(ci,e k ) if i ∈ N C , j ∈ N E (<label>8</label></formula><formula xml:id="formula_9">)</formula><p>2 http://www.cs.waikato.ac.nz/ml/weka</p><p>where N E (i) is the set of entity nodes such that for each node k ∈ N E (i), there is an edge from i to k in G. The entry v i in v is the initial evidence representing the prior importance of a mention m i if i ∈ N M or the document-specific importance of an category c i if i ∈ N C , which is calculated as follows</p><formula xml:id="formula_10">v i =        λ•P (mi) λ• k∈N M P (m k )+η• k∈N C P (c k ) if i ∈ N M η•P (ci) λ• k∈N M P (m k )+η• k∈N C P (c k ) if i ∈ N C 0 otherwise<label>(9)</label></formula><p>where λ and η are tunable parameters with λ+η = 1, which reflect the sensitivity of prior mention importance and document-specific category importance to the final probability of each candidate entity. When η = 0, our method reduces to general entity linking without considering the topic-sensitive model. In contrast, when λ = 0, the initial evidence of the graph-based disambiguation only depends on the category importance. As a result of the personalized PageRank algorithm, each candidate entity e receives a final probability P (e). For each mention m having a set of candidate entities E m , we choose the entity with the maximal probability as the predicted linking entity, i.e., e m = arg max e∈Em P (e). The process discussed above doesn't distinguish between salient and non-salient entities. In order to deal with salient entity linking, one important task of the topic-sensitive model is to validate whether the predicted linking entity e m for mention m is a salient entity. For this purpose, we learn a threshold τ such that if P (e m ) is greater than τ we return e m as the linking entity for m, otherwise we return Non-Salient.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments</head><p>We now discuss the experiments we performed to assess the performance of our approach. As the knowledge base, we used the English Wikipedia snapshot from July 2013. We employed the Reuters-128 entity salience dataset<ref type="foot" target="#foot_1">3</ref> , which is an extension of a part of the N3 entity linking datasets <ref type="bibr" target="#b16">[17]</ref>. The Reuters-128 dataset is an English corpus and it contains 128 economic news articles. The dataset contains information for 880 named entities with their position in the document and a URI of a DBpedia resource identifying each entity. The salience dataset extends the Reuters-128 dataset also with 3,551 common entities.</p><p>In order to construct the dataset, entity salience information was obtained by crowdsourcing salience information using the CrowdFlower platform. For each named and common entity in the Reuters-128 dataset, the authors of the dataset collected at least three judgements. Only judgments from annotator with trust score higher than 70% were considered as trusted judgements. If the trust score of an annotator falls bellow 70%, all his/her judgements were disregarded. Finally, each named and common entity in the dataset has been classified in one of the following classes -Most Salient -Entities with the highest focus of attention in the article. The document is mostly about the these entities, or the entities play a prominent role in the content of the article. -Less Salient -Entities with less focus of attention in the article. The entities play an important role in some parts of the content of the article. -Not Salient -The article is really not about the entities</p><p>In our experiments, we consider the entities in both classes Most Salient and Less Salient as salient entities, while entitites belonging to Not Salient are considered as non-salient entities. Using the Reuters-128 entity salience dataset, we conducted the experiments to compare our approach with several entity linking methods. We used two variants of our approach, one employs only the graph-based disambiguation for general entity linking (λ = 1 and η = 0) and the other integrates the topic-sensitive model with the goal of salient entity linking (λ = 0.2 and η = 0.8). All the methods should label each mention with either the correct entity or Not Salient. Note that we restrict the input to the labeled mentions to compare the method's ability to distinguish between salient entity and non-salient entity, not its ability to recognize entity names in the input text. The adopted evaluation criteria include Micro-Precision, Micro-Recall, Micro-F1, Macro-Precision, Macro-Recall and Macro-F1.</p><p>The experimental results are shown in Table <ref type="table" target="#tab_1">1</ref>. By utilizing the topic-sensitive model, our approach to salient entity disambiguation significantly outperforms the baselines in terms of all evaluation criteria. Regarding the two variants of our approach, it clearly shows that the topic-sensitive model indeed contributes to the final performance improvement.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>In this paper, we introduce the task of salient entity linking that existing entity linking solutions cannot well address. For tackling this new problem, we propose a graph-based disambiguation framework, which integrates several features including prior mention importance, mention-entity compatibility, entity-entity coherence and in particular a topic-sensitive model capturing entity-category association and document-specific category importance. We have experimentally shown that our approach achieves a significant improvement over the baselines. The evaluation results also show that the topic-sensitive model indeed helps with the salient entity disambiguation.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 :</head><label>1</label><figDesc>Fig. 1: Salient entity linking framework.</figDesc><graphic coords="2,159.04,115.83,293.96,166.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>s as anchor text pointing to e as destination and E m.s is the set of entities that have the surface form m.s. An entity e is characterized by its textual description e.c, called context of e and a mention m is characterized by its surrounding sentences m.c, called context of m. The context similarity CS(m, e) between m and e can be calculated using cosine similarity on the term vectors e.c of e.c and m.c of m.c as CS(m, e) = cos(e.c, m.c) = e.c, m.c |e.c| • |m.c|</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>4 :</head><label>4</label><figDesc>MethodsMic. Prec. Mic. Rec. Mic. F1 Mac. Prec. Mac. Rec. Mac. F1.</figDesc><table><row><cell>DBpedia Spotlight [2]</cell><cell>0.45</cell><cell>0.39</cell><cell>0.41</cell><cell>0.45</cell><cell>0.37</cell><cell>0.40</cell></row><row><cell>Wikipedia Miner [1]</cell><cell>0.60</cell><cell>0.48</cell><cell>0.54</cell><cell>0.60</cell><cell>0.42</cell><cell>0.52</cell></row><row><cell>NERD-ML [5,7]</cell><cell>0.67</cell><cell>0.50</cell><cell>0.57</cell><cell>0.65</cell><cell>0.46</cell><cell>0.54</cell></row><row><cell>WAT [4,8]</cell><cell>0.35</cell><cell>0.32</cell><cell>0.34</cell><cell>0.36</cell><cell>0.33</cell><cell>0.34</cell></row><row><cell>AGDISTIS [6]</cell><cell>0.73</cell><cell>0.50</cell><cell>0.59</cell><cell>0.73</cell><cell>0.48</cell><cell>0.58</cell></row><row><cell>Our Method (General)</cell><cell>0.70</cell><cell>0.46</cell><cell>0.56</cell><cell>0.69</cell><cell>0.45</cell><cell>0.55</cell></row><row><cell>Our Method (Salient)</cell><cell>0.83</cell><cell>0.51</cell><cell>0.63</cell><cell>0.82</cell><cell>0.50</cell><cell>0.62</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>The experimental results.</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">In this work, we employ the 16 second-level categories including Mathematics, People, Science, Sport, Geography, Culture, Politics, Nature, Technology, Education, Health, Business, Belief, Society, Life and Concepts in Wikipedia, where the first-level category is the fundamental category.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://github.com/KIZI/ner-eval-collection</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">http://ner.vse.cz/datasets/entitysalience-collection</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgments. The research leading to these results has received funding from the European Union Seventh Framework Programme (FP7/2007-2013) under grant agreement no. 611346.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Learning to link with wikipedia</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">N</forename><surname>Milne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIKM</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="509" to="518" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Dbpedia spotlight: shedding light on the web of documents</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">N</forename><surname>Mendes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakob</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>García-Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">I-SEMANTICS</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Collective entity linking in web text: a graph-based method</title>
		<author>
			<persName><forename type="first">X</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="765" to="774" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Fast and accurate annotation of short texts with wikipedia pages</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ferragina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Scaiella</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Software</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="70" to="75" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Learning with the web: Spotting named entities on the intersection of NERD and machine learning</title>
		<author>
			<persName><forename type="first">M</forename><surname>Van Erp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Rizzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Troncy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">#MSM</title>
		<imprint>
			<biblScope unit="page" from="27" to="30" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">AGDISTIS -graph-based disambiguation of named entities using linked data</title>
		<author>
			<persName><forename type="first">R</forename><surname>Usbeck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Ngomo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Röder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gerber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Coelho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Both</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ISWC</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="457" to="471" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Benchmarking the extraction and disambiguation of named entities on the semantic web</title>
		<author>
			<persName><forename type="first">G</forename><surname>Rizzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Van Erp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Troncy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LREC</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="4593" to="4600" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">From tagme to WAT: a new entity annotator</title>
		<author>
			<persName><forename type="first">F</forename><surname>Piccinno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ferragina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ERD@SIGIR</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="55" to="62" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Identifying salient entities in web pages</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gamon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Apacible</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pantel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIKM</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="2375" to="2380" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">An effective, low-cost measure of semantic relatedness obtained from wikipedia links</title>
		<author>
			<persName><forename type="first">I</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Milne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">WIKIAI</title>
		<imprint>
			<biblScope unit="page" from="25" to="30" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The google similarity distance</title>
		<author>
			<persName><forename type="first">R</forename><surname>Cilibrasi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">M B</forename><surname>Vitányi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Knowl. Data Eng</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="370" to="383" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Advances in kernel methods</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Platt</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>MIT Press</publisher>
			<biblScope unit="page" from="185" to="208" />
			<pubPlace>Cambridge, MA, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Improvements to platt&apos;s smo algorithm for svm classifier design</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Keerthi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Shevade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bhattacharyya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R K</forename><surname>Murthy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Comput</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="637" to="649" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Classification by pairwise coupling</title>
		<author>
			<persName><forename type="first">T</forename><surname>Hastie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tibshirani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NIPS</title>
				<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="507" to="513" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Scaling personalized web search</title>
		<author>
			<persName><forename type="first">G</forename><surname>Jeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Widom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">WWW</title>
		<imprint>
			<biblScope unit="page" from="271" to="279" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Topic-sensitive pagerank: A context-sensitive ranking algorithm for web search</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">H</forename><surname>Haveliwala</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Knowl. Data Eng</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="784" to="796" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">N 3 -A collection of datasets for named entity recognition and disambiguation in the NLP interchange format</title>
		<author>
			<persName><forename type="first">M</forename><surname>Röder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Usbeck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hellmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gerber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Both</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LREC</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="3529" to="3533" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
