<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Background Linking: Joining Entity Linking with Learning to Rank Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Ornella</forename><surname>Irrera</surname></persName>
							<email>irreraorne@dei.unipd.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Information Engineering</orgName>
								<orgName type="institution">University of Padova</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gianmaria</forename><surname>Silvello</surname></persName>
							<email>silvello@dei.unipd.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Information Engineering</orgName>
								<orgName type="institution">University of Padova</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Background Linking: Joining Entity Linking with Learning to Rank Models</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">9F2144B0B0A8700EF53247A8185FF868</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T16:13+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Entity Linking</term>
					<term>Graph of Entities</term>
					<term>Learning to Rank</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The recent years have been characterized by a strong democratization of news production on the web. In this scenario it is rare to find self-contained news articles that provide useful background and context information. The problem of finding information providing context to news articles has been tackled by the Background Linking task of the TREC News Track. In this paper, we propose a system to address the background linking task. Our system relies on LambdaMART learning to rank algorithm trained on classic textual features and on entity-based features. The idea is that the entities extracted from the documents as well as their relationships provide valuable context to the documents. We analyzed how this idea can be used to improve the effectiveness of (re-)ranking methods for the background linking task.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>According to Pew Research studies carried out in 2018, the 93% of American adults consume at least some of their news online, via social media recommendations, web browsing or advertising recommendations <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b21">22]</ref>. The adoption of digital strategies marked the end of the publisher-driven news delivery, shifting the focus faraway from the publisher towards the story. In this scenario, the user is allowed to consume news on the web and publish his ones. This had a substantial impact on news production. Indeed, it is more and more challenging to find self-contained news articles and provide context and background information about the story told. The National Institute of Standards and Technology (NIST) recognized to Information Retrieval (IR) and Natural Language Processing (NLP) a primary role in the solution of this problem and, in cooperation with the Washington Post, launched the first edition of the News Track in TREC 2018 <ref type="bibr" target="#b13">[14]</ref>. This track is organized into two subtasks, Background Linking and Entity Ranking: their goal is to provide the user with different means to understand news articles. The former relies on the construction of a ranked list of articles to provide background information. The latter exploits, as a means of contextualization, a ranking of named entities mentioned in the article the user is reading.</p><p>In this paper, we present an IR system based on Learning to Rank methods to improve the ranking of documents for the Background Linking task context.</p><p>In the literature, several solutions were proposed to address this task. Most of them treat background linking as an ad hoc search task and rely on approaches based on keyword extraction <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b25">26]</ref>. Other methods instead, leverage on entities to identify documents' topics <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>. In the 2019 edition of the TREC Background Linking task instead, many participants exploited machine learning methods <ref type="bibr" target="#b6">[7]</ref>, with a particular focus on learning to rank approaches <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b19">20]</ref>.</p><p>We propose a retrieval system that relies on LambdaMART learning to rank algorithm <ref type="bibr" target="#b2">[3]</ref> and the classic BM25 model <ref type="bibr" target="#b20">[21]</ref> to rank background articles. In particular, we focus on the creation of feature vectors to be fed to learning to rank methods such as LambdaMART; indeed, we combine features extracted from two different representations of the same document: the unstructured textual representation and a graph-based one. We consider a graph of entities extracted from the textual documents and then linked to Wikipedia articles. In this paper, we analyze advantages and limits of entity-based features in learning to rank approaches and we discuss how they can be employed to improve the final document ranking for the background linking task.</p><p>The rest of the article is organized as follows: in Section 2 we provide some background and related work, in Section 3 we describe the key components of the proposed solution and in Section 4 we present a use case. Section 5 describes the experimental setup and Section 6 reports about the evaluation results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Background</head><p>The TREC News Track aims to study how to provide contextual information to users while reading news articles. To this end, two tasks are defined: Entity Ranking and Background Linking. Background Linking Task concerns the development of systems able to help users contextualize news articles as they are reading them. Formally, given a source article (i.e., the query), the system should retrieve a list of articles providing relevant background and context information related to the source <ref type="bibr" target="#b13">[14]</ref>.</p><p>The reference collection for this task is the Washington Post Corpus ver.2.<ref type="foot" target="#foot_0">1</ref> This collection contains about 590,000 news articles and blog posts published from 2012 to 2017 by the Washington Post<ref type="foot" target="#foot_1">2</ref> . Each document is characterized by a list of fields such as id, author, article URL, date of publication, title. The main content is organized in one or more paragraphs; they may include HTML tags, images and videos. Fifty topics have been provided for this task; each of them is identified by a number and by the id and the URL to the topic's source document  (or query). Graded relevance judgments are ranging from 0 (not relevant) to 4 (recommended article).</p><p>In the proposed solution, we leverage document graph representation based on the entities and their relationships that we automatically extract from the documents. In the literature, several approaches take advantage of entity-oriented representations. Xiong et al. in <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b24">25]</ref> propose a Bag-of-Entities (BoE) model where each document is represented as a bag-of-entities constructed via entity linking. The ranking of documents is generated, considering the overlapping entities between each document and the query. A similar approach is <ref type="bibr" target="#b10">[11]</ref> that describes a learning to rank approach where the training is based both on Bagof-Words (BoW) and BoE features. The experimental evaluation showed that the combination of features depending on words and entities improve a BM25 baseline. What makes our retrieval system different from the solutions proposed above, is that we take the separated entities belonging to the BoE constructed via entity linking and we create a graph; this graph is more informative than the classic BoE representation because it includes both the information related to the separated entities and the information carried by their relationships.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Proposed Solution</head><p>The main phases composing the proposed solution are reported in Figure <ref type="figure" target="#fig_0">1</ref>. In the following, we describe every single phase.</p><p>Entity Linking. In this phase, we perform entity recognition on the textual documents to extract a set of mentions to be linked with entities in a knowledge repository (KR) <ref type="bibr" target="#b0">[1]</ref> -i.e., Wikipedia<ref type="foot" target="#foot_2">3</ref> in our case. This process can be called entity annotation. This phase's output is a bag-of-entities for each document.</p><p>Graph Creation. This phase takes the bag-of-entities as input and creates an undirected weighted graph, where the nodes are the entities and the edges are based on the semantic relatedness between the nodes. The semantic relatedness is a measure defined in <ref type="bibr" target="#b22">[23]</ref>, which exploits the Wikipedia structure to find the relatedness between two entities. Two entities are semantically related if they share a high number of entities linking to them <ref type="bibr" target="#b0">[1]</ref>. In our implementation, the entity pairs are connected by an edge whose weight is the numerical value of the semantic relatedness. The output of this phase is a graph with one or more connected components. There could be a strong imbalance between the different components; in fact, one may include the largest part of the nodes, while another contains very few entities. It is common to identify in the largest connected component one or more communities -or groups of nodes highly connected within themselves and poorly with the other groups <ref type="bibr" target="#b3">[4]</ref>.</p><p>Graph Pruning. In this phase we remove the meaningless entities from the graph. Pruning is based on components removal, which keeps only the largest connected component of the graph and community detection that detects the largest community, where the most representative entities usually are. To this end, we rely on the Girvan-Neman algorithm for community detection <ref type="bibr" target="#b9">[10]</ref> based on the "edge betweenness", a generalized version of the classic betweenness centrality measure defined in <ref type="bibr" target="#b7">[8]</ref>. The edge betweenness of an edge corresponds to the number of shortest paths between every pair of vertices that run along it <ref type="bibr" target="#b9">[10]</ref>. Removing some nodes from a graph may cause a loss of information, especially if two large connected components coexist in the same graph (in this case the separation between different components shows that the article discusses two weakly correlated subjects). Since this case is extremely rare, the advantages brought by the pruning phase predominate over the possible loss of information.</p><p>Features Extraction. This phase is about the definition and extraction of the features representing the documents; there are document-based and query-based features. The two feature sets comprise both textual and entity-based graph features. Since the overall goal of the implementation consists in studying the impact of the graphs of entities in the learning process, the core of this phase lays in the extraction of the entity-based graph features. The query graph-based features are the most informative about the relevance of a document because they reflect the similarities between the document graph and the query one. This is why the largest part of the extracted features belong to this type. An example of query graph-based features is the semantic relatedness <ref type="bibr" target="#b22">[23]</ref> between the most central node of each graph, where the centrality measure considered is the betweenness centrality <ref type="bibr" target="#b7">[8]</ref>. High values highlight the correlation between the document graph and the query one. The features computed independently of the query graph, instead, are topological properties of the document graph. This type of features is less informative than the previous one because it does not highlight any relation of the document with the query; however properties such as the node connectivity and the node's degree revealed their usefulness in relevance identification. Only few features belong to the textual-based type, and most of them depend on the query article. Some examples of query-based textual features are the BM25 score and the Term Frequency (TF), while the document's length and the number of paragraphs are document-based textual features. In Table <ref type="table" target="#tab_1">1</ref> it is presented an overview of the number of features extracted from each document's representation. For each document it is finally created a feature vector. These vectors are required to perform learning to rank tasks.</p><p>Learning to Rank. We employ LambdaMART because it is one of the best performing learning to rank methods <ref type="bibr" target="#b2">[3]</ref> in the literature. LambdaMART is a list-wise approach based on Multiple Additive Regression Trees (MART) <ref type="bibr" target="#b8">[9]</ref>.</p><p>LambdaMART is trained to automatically construct a set of ranking models.</p><p>The ranking models trained on different sets of hyper-parameters are tested on the test set containing a new list of queries; the documents in the test set are ranked on the base of the new scores assigned by the models. The output of this phase is the set of final runs, one for each model.</p><p>Fusion. In this phase we experiment the fusion of multiple system's runs to analyze whether an increase in the number of merged runs corresponds to an effectiveness improvement. Each run provided by the learning to rank phase contains the same set of documents ranked differently. In order to create the fusion runs we employ combSUM method <ref type="bibr" target="#b17">[18]</ref>: the new final score of a document corresponds to the sum of the scores that document obtained in each individual run. The documents are finally re-ranked according to their new scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Use case</head><p>To better illustrate how a graph of entities is created starting from a textual article, we employ a small portion of an article<ref type="foot" target="#foot_3">4</ref> about tropical storms.</p><p>Entity Linking. The mentions are detected and linked to Wikipedia entities. Below, we report the textual fragment where the mentions are marked in boldface. Fig. <ref type="figure">2</ref>: An example of entity graph. In red the nodes and edges removed after the graph pruning phase.</p><p>In Table <ref type="table" target="#tab_3">2</ref> we report the mentions and their linked entities. Giving a first look to the mention-entity associations, the largest part of entities are coherent with the tropical storms except for past tense and 24-hour clock which are very general or out of context. Graph Creation. The extracted entities are the nodes of the graph, while the edges are based on the semantic relatedness between the nodes. In Figure <ref type="figure">2</ref> we show the resulting graph. The two connected components are highly unbalanced because a component includes the largest part of entities while the second has only two entities. In the largest connected component lie the most meaningful entities, while the smallest one contains the two out of context entities, past tense and 24-hour clock.</p><p>Graph Pruning. Starting from the graph in Figure <ref type="figure">2</ref>, we firstly remove the smallest connected component (right part in red). Then, we run the Girvan-Newman algorithm that removes the Earth and Explosion entities. These entities are discarded because they have degree equal to one and they do not belong to the largest community. Nonetheless, if we consider these entities in the context of tropical storms, they do not bring any contribution for context identification, hence their removal increases the coherence of the graph with respect the topic.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experimental Setup</head><p>In this section, we describe the experimental setup. In particular, we discuss some technical details, and we illustrate the tools used to implement each phase of the proposed solution.</p><p>Preprocessing. We preprocessed the Washington Post Corpus to remove from each article all the fields without informative content. Then we indexed the collection with Apache Solr <ref type="foot" target="#foot_4">5</ref> , an open-source search library based on Apache Lucene set with the default English tokenizer, the English stop-words list and no stemmer. Once indexed the collection, we used the default BM25 in Solr to construct an initial ranking of 100 documents per topic: our baseline.</p><p>Entity Linking and Graph creation. Since it was unfeasible to analyze the entire collection of documents in order to choose the most effective settings' apporach, we relied on a sample of thirty documents, both relevant and not relevant, and we considered the setting the most effective on this sample. The linking system we adopted is TagMe <ref type="bibr" target="#b5">[6]</ref>, in particular, we performed article annotation via the TagMe RESTful API, setting a confidence score ρ = 0.1. This parameter is a threshold imposed to discard non-meaningful entities <ref type="bibr" target="#b12">[13]</ref>: the lower this parameter is, the more entities are recognized. For each article, we linked at most thirty entities leading to graphs with at most thirty nodes. To create graphs as consistent as possible with the original textual article, we set a threshold equal to 0.4 on the value of semantic relatedness. Specifically, if two entities have a semantic relatedness higher (or equivalent) than 0.4, an edge between them is created. This threshold allowed us to obtain graphs subdivided into connected components: such a structure highlights the distinction between meaningful and non-meaningful entities. Pruning is performed only if the graph contains at least ten nodes: if it does not, pruning will lead to a graph unable to represent the original document correctly. To perform graph creation and pruning, we relied on NetworkX Python library <ref type="bibr" target="#b11">[12]</ref>.</p><p>Features Extraction. To extract the set of textual-based features from each document, we exploited Scikit-learn <ref type="bibr" target="#b18">[19]</ref>, a Python library that provides the implementation of measures like the TF and the Inverse Document Frequency (IDF).</p><p>For the extraction of the entity-based graph features, both document-based and query-based, we exploited the NetworkX Python library <ref type="bibr" target="#b11">[12]</ref>. We created for each document a feature vector containing the following elements: the relevance judgment of the document, the id of the query, the features extracted and the id of the document.</p><p>Learning to Rank. We relied on the implementation of LambdaMART provided by RankLib<ref type="foot" target="#foot_5">6</ref> , a library of learning to rank algorithms developed by the Lemur Project. We performed manual tuning, which allowed us to understand the best combinations of hyperparameters and analyze how each parameter interferes with the others. Table <ref type="table" target="#tab_4">3</ref> summarizes the main features related to the training and test sets involved in our implementation. The test set comprises the vector representation of the documents belonging to the baseline. The training set, instead, combines two sets of topics belonging to different TREC tracks. We remark that even if both the training and test sets contain documents belonging to the Washington Post Corpus, they rely on disjointed sets of topics. Performing training depending on two different collections of documents allowed us to enrich the training set but, at the same time, it made our system suffer from the limitations induced by the transfer learning. In particular, what influenced our system the most was the lack of consistency between the relevance grades provided for the two tracks; indeed, we had to map the five possible relevance grades offered for the Background Linking task to the three grades provided for the Common Core Track of TREC 2017. The training set was finally split to derive a validation set. We trained the algorithm by choosing more than 200 different combinations of hyperparameters, and we selected the seventy models which maximized the nDCG@5 effectiveness on the validation set, and, at the same time, which were far from overfitting or underfitting conditions. In Table <ref type="table" target="#tab_5">4</ref>, the values of the five most effective models' hyperparameters are described. We tested the seventy ranking models on the test set, and we obtained seventy final runs: each one contains a different re-ranking of the baseline documents.</p><p>Fusion. To perform fusion we considered the ten most effective models and the related final runs. We fused k (with k = {2, 3, 5, 7, 10}) runs, and, for each k, we collected ! 10 k " runs -i.e. all the possible combinations of k distinct runs taken from a set of ten runs. Evaluation and metrics. Coherently with the guidelines of the task <ref type="bibr" target="#b13">[14]</ref>, we considered nDCG@5 as primary measure of effectiveness and we tested the system's performances also in terms of nDCG@1, nDCG@10, nDCG@100, Precision at Cut-off 1 (P@1) and reciprocal rank (recip rank). We analyzed two types of runs:</p><p>(1) the individual run, intended as the re-ranking produced by each one of the seventy LambdaMART's models, and (2) the fused run, intended as the result of the fusion of k individual runs belonging to the set of ten runs dedicated to the fusion approach. We used trec eval<ref type="foot" target="#foot_6">7</ref> tool to evaluate the set of seventy individual runs and the five sets of fused runs (one set of fused runs for each k); for each set we selected the run with the highest nDCG@5, obtaining in total six selected runs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Evaluation and results</head><p>In this section we describe the results obtained by our system. We assess the retrieval effectiveness from both a quantitative and a qualitative viewpoint.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Quantitative evaluation</head><p>The quantitative evaluation has the intent to describe how our system performs on average. To this end, in Table <ref type="table" target="#tab_6">5</ref> and Table <ref type="table" target="#tab_7">6</ref> we illustrate the average results of the six selected runs and the baseline. We indicated the baseline as BM25, the best individual run as best run and the five best runs obtained with the fusion approach as fus followed by the number of runs fused. The top scores of each evaluation measure are highlighted in boldface.</p><p>In Table <ref type="table" target="#tab_6">5</ref> we describe the performances of our system in terms of reciprocal rank, P@1 and nDCG@1: these metrics evaluate the effectiveness of one document per topic in a run. Our goal is to evaluate if our system can correctly recognize a relevant document and place it in the ranking's highest positions. In general, all the generated runs show effectiveness slightly higher than the baseline both for reciprocal rank and nDCG@1. In P@1 instead, three runs equalled the baseline (best run, fus2 and fus10 ) and three runs reported greater values (fus3, fus5 and fus7 ). The most effective run is fus5, obtained fusing five individual runs. It maximizes both reciprocal rank and P@1 evaluation measures. best run and fus7 instead, maximize the nDCG@1. Observing the reported results, we can notice that increasing the number of runs fused does not necessarily correspond to an effectiveness improvement. The effectiveness of fus5 and fus7 runs in fact, is usually higher than the effectiveness of fus2, fus3 and fus10 : this indicates that the fusion can improve the performances until a certain number of runs fused, after that, the fusion becomes disadvantageous and the effectiveness decreases. This is proved by fus10 which is the least effective run among the five fused runs. This behavior is verified in all the measures proposed in Table <ref type="table" target="#tab_6">5</ref>. Table <ref type="table" target="#tab_7">6</ref> reports the effectiveness of the proposed runs at different ranking depths. This analysis considers the nDCG evaluation measure at different cutoffs: 5, 10 and 100 (in this case, it is evaluated the entire ranking associated with each topic). In this analysis, we study our system's ability to produce effective rankings. In particular, we are interested in how the effectiveness varies, increasing the cut-off. The results reported in terms of nDCG@5 highlight that the fus2, fus3, fus5 and fus7 runs have performances slightly above those of the baseline, while the best run approximates it very well. The least effective run in nDCG@5 is the fus10 which achieves the lowest result. If we consider more than five documents in the ranking, the BM25 model always reports the Fig. <ref type="figure">3</ref>: This histogram represents the nDCG@5 effectiveness of the best run compared to the nDCG@5 of the baseline. Each column represents a topic and the column's height is found computing the difference between the nDCG@5 of the best run and the nDCG@5 of the baseline for that topic. highest effectiveness. The highest results achieved by our system for nDCG@10 and nDCG@100 are those of fus5, while the lowest ones are attributed to the best run and fus10. As it has been already noticed in Table <ref type="table" target="#tab_6">5</ref>, this revealed that the fusion approach improves the system's effectiveness, but it becomes disadvantageous if more than five runs are fused.</p><p>Discussion First, our hypothesis that fusing more runs leads to an effectiveness improvement does not seem to be supported. Both in Table <ref type="table" target="#tab_6">5</ref> and Table <ref type="table" target="#tab_7">6</ref> the fusion of all the available runs is the least effective run among the six proposed. We also see that there is no correlation between the number of runs fused and the effectiveness improvement. Despite this fact, the fusion approach has been revealed to be useful in our implementation, since among the selected runs, the fused ones achieve the highest results. Moreover, our system and the BM25 model show two opposite behaviors: the more documents we consider in the ranking, the more our system's effectiveness decreases, and vice versa holds for the BM25 model. This contrast between our system and BM25 is primarily caused by the transfer learning. In fact, the effectiveness decrease detected after only five documents is strictly related to the inability of our system to distinguish among more than three relevance scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Qualitative evaluation</head><p>The qualitative evaluation describes the effectiveness of the particular topics belonging to a run. The histogram shown in Figure <ref type="figure">3</ref> describes the difference between the nDCG@5 score of the best run for a given topic and that one of the baseline. In particular, all the columns above the abscissa identify the topics where our system prevails over the baseline. Vice versa holds for the columns below the abscissa. It is possible to notice an equivalent number of topics laying above and below the abscissa; this means that the effectiveness improvement brought by the topics where our system prevails over the baseline is perfectly balanced by the effectiveness decrease caused by the topics where BM25 prevails over our system. This is the reason why, in the quantitative analysis, the nDCG@5 of the best run (0.4090) approximated the one of the baseline (0.4097).</p><p>We conducted an analysis focused on the rankings associated with the best and the worst performing topics of the best run -this study aimed at finding a correlation between the documents' feature vectors and our system's effectiveness. The most compelling topics report in the first positions of their rankings documents whose relevance is identifiable both in the textual-based and in the entity-based graph features. This proves that relying on features that leverage the classic textual representation and the graph of entities can lead to an effectiveness improvement. Analyzing the rankings associated with the topics where the BM25 prevails, we identified three main reasons for our system's poor effectiveness in some topics. These are:</p><p>-The transfer learning: it prevented some topics from reaching a high level of effectiveness. -The graphs of entities: the documents may not present well-formed graphs; in this case, the learner attributes a score on the basis of the textual features only. This happens when the original article has not enough textual content to construct a consistent graph. -Errors in learning phase: this condition depends on the features in the vectors; we experimented that the value of a single feature may influence the entire ranking of a topic.</p><p>These conditions are unavoidable, and they represent the cases in which it is convenient to use the BM25 model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions</head><p>This article presented a solution for the background linking task relying on LambdaMART to obtain a list of background articles. We leveraged the document's textual and graph representations to extract a set of features used to perform training. Our goal was to study whether the combination of features belonging to different document representations can improve the system's effectiveness; in particular, we were interested in exploring the impact related to the graphs of entities-based features. The analysis conducted on the single topics highlighted that there is a set of topics where our system outperformed the BM25 model. In these cases, the graphs of entities played a crucial role because the combination of textual-based features and graph-based ones allowed for an effectiveness improvement. This implied that our initial hypothesis is confirmed for this first set of topics. However, there was an equivalent number of topics where our system was not highly effective. Transfer learning has a high impact on negative performances. The balance between the effective and ineffective topics explained the similarity between the average nDCG@5 values obtained by our system and by the baseline. We finally introduced the fusion approach as a means to improve the overall effectiveness of our system. The results showed that, contrary to our initial belief, fusing too many runs makes the performances decrease; in particular, the optimal number of runs to merge, in the tested context, is five.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 :</head><label>1</label><figDesc>Fig. 1: Representation of the main phases composing the proposed solution.Green rectangles show the six phases, while the blue ones show the output of each phase.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="11,121.89,115.92,368.50,170.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>1 Entit Linking 2 Graph Creation 3 Graph Pruning 4 Features Extraction Learning to Rank Fusion</head><label></label><figDesc></figDesc><table><row><cell>Textual documents</cell><cell></cell><cell></cell><cell></cell><cell>Final Run</cell></row><row><cell>Bag of Entities</cell><cell>Initial Graph</cell><cell>Pruned Graph</cell><cell>Features Vectors</cell><cell>Re-rankings of test documents</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Overview of the features extracted.</figDesc><table><row><cell>Type</cell><cell>Subtype</cell><cell cols="3"># of features Total</cell></row><row><cell>Textual-based features</cell><cell>Document-based Query-based</cell><cell>3 7</cell><cell>10</cell><cell>.</cell></row><row><cell>Entity-based graph features</cell><cell>Document graph-based Query graph-based</cell><cell>12 43</cell><cell>55</cell><cell>.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>Super Typhoon Vongfong explodes becomes most intense storm on Earth in 2014. Super Typhoon Vongfong has rapidly intensified over the past 24 hours from the equivalent of a category two hurricane to a monster typhoon[...]   </figDesc><table><row><cell cols="2">Earth</cell><cell>Tropical cyclone scales</cell><cell>0.560</cell><cell>Typhoon Vongfong</cell><cell>Past tense</cell></row><row><cell>0.450</cell><cell>0.483</cell><cell>0.963</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>0.560</cell><cell></cell></row><row><cell>Storm</cell><cell></cell><cell></cell><cell></cell><cell>0.612</cell></row><row><cell></cell><cell>0.514</cell><cell>0.850</cell><cell></cell><cell></cell><cell>0.455</cell></row><row><cell>0.405</cell><cell>0.648</cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Saffir-Simpson</cell></row><row><cell>Explosion</cell><cell></cell><cell>Typhoon cyclone</cell><cell>0.877</cell><cell>hurricane Wind scale</cell><cell>24-hour clock</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2 :</head><label>2</label><figDesc>Mention -Wikipedia entity associations</figDesc><table><row><cell>Mention</cell><cell>Wikipedia entity</cell></row><row><cell>Super Typhoon</cell><cell>Tropical cyclone scales</cell></row><row><cell>explodes</cell><cell>Explosion</cell></row><row><cell>storm</cell><cell>Storm</cell></row><row><cell>Earth</cell><cell>Earth</cell></row><row><cell>Typhoon Vongfong</cell><cell>Typhoon Vongfong (2014)</cell></row><row><cell>past</cell><cell>Past tense</cell></row><row><cell>24 hours</cell><cell>24-hour clock</cell></row><row><cell>category two hurricane</cell><cell>SaffirSimpson hurricane wind scale</cell></row><row><cell>typhoon</cell><cell>Tropical cyclone</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 3 :</head><label>3</label><figDesc>Overview of training and test sets.</figDesc><table><row><cell>Set</cell><cell>Track</cell><cell>Collection</cell><cell>Topics</cell><cell># of Vectors</cell><cell>Total</cell></row><row><cell>Training set</cell><cell>News Track 2019 Background Linking task Common Core Track 2017</cell><cell>Washington Post Corpus vs2 New York Times Annotated Corpus</cell><cell>57 44</cell><cell>8127 12736</cell><cell>20863</cell></row><row><cell>Test set</cell><cell>News Track 2018 Background Linking task</cell><cell>Washington Post Corpus vs2</cell><cell>50</cell><cell>5000</cell><cell>5000</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 4 :</head><label>4</label><figDesc>Hyperparameters' combinations of the six best models. In bold text we marked the combination related to the model whose nDCG@5 was the highest.</figDesc><table><row><cell>Trees</cell><cell cols="3">Leaves Shrinkage Threshold candidates</cell></row><row><cell>600</cell><cell>6</cell><cell>0.08</cell><cell>all</cell></row><row><cell>1000</cell><cell>7</cell><cell>0.06</cell><cell>256</cell></row><row><cell>2000</cell><cell>5</cell><cell>0.05</cell><cell>all</cell></row><row><cell>2000</cell><cell>8</cell><cell>0.03</cell><cell>256</cell></row><row><cell>2500</cell><cell>5</cell><cell>0.01</cell><cell>all</cell></row><row><cell>2500</cell><cell>8</cell><cell>0.01</cell><cell>256</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 5 :</head><label>5</label><figDesc>Evaluation results of one document per topic.</figDesc><table><row><cell></cell><cell>recip rank</cell><cell>P@1</cell><cell>nDCG@1</cell></row><row><cell>best run</cell><cell>0.8128</cell><cell>0.7400</cell><cell>0.4150</cell></row><row><cell>fus2</cell><cell>0.8269</cell><cell>0.7400</cell><cell>0.3825</cell></row><row><cell>fus3</cell><cell>0.8378</cell><cell>0.7600</cell><cell>0.4100</cell></row><row><cell>fus5</cell><cell>0.8473</cell><cell>0.7800</cell><cell>0.4100</cell></row><row><cell>fus7</cell><cell>0.8373</cell><cell>0.7600</cell><cell>0.4150</cell></row><row><cell>fus10</cell><cell>0.8240</cell><cell>0.7400</cell><cell>0.3950</cell></row><row><cell>BM25</cell><cell>0.7951</cell><cell>0.7400</cell><cell>0.3525</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 6 :</head><label>6</label><figDesc>Evaluation results at different cut-offs.</figDesc><table><row><cell></cell><cell cols="3">nDCG@5 nDCG@10 nDCG@100</cell></row><row><cell>best run</cell><cell>0.4090</cell><cell>0.4055</cell><cell>0.4589</cell></row><row><cell>fus2</cell><cell>0.4168</cell><cell>0.4100</cell><cell>0.4601</cell></row><row><cell>fus3</cell><cell>0.4166</cell><cell>0.4106</cell><cell>0.4635</cell></row><row><cell>fus5</cell><cell>0.4129</cell><cell>0.4131</cell><cell>0.4646</cell></row><row><cell>fus7</cell><cell>0.4139</cell><cell>0.4038</cell><cell>0.4636</cell></row><row><cell>fus10</cell><cell>0.4017</cell><cell>0.4077</cell><cell>0.4602</cell></row><row><cell>BM25</cell><cell>0.4097</cell><cell>0.4233</cell><cell>0.4659</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://trec.nist.gov/data/wapost/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://www.washingtonpost.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://www.wikipedia.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">Article available at: https://www.washingtonpost.com/news/capital-weathergang/wp/2014/10/07/super-typhoon-vongfong-explodes-becomes-most-intensestorm-on-earth-in-2014/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://lucene.apache.org/solr/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://sourceforge.net/p/lemur/wiki/RankLib/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://trec.nist.gov/trec eval/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Entity-oriented search</title>
		<author>
			<persName><forename type="first">K</forename><surname>Balog</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>Springer Nature</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">htw saar@ trec 2018 news track</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bimantara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Engelhardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gerwert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gottschalk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lukosz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Piri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">S</forename><surname>Shaft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Berberich</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">From ranknet to lambdarank to lambdamart: An overview</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Burges</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Learning</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page">81</biblScope>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Community structure in networks: Girvan-newman algorithm improvement</title>
		<author>
			<persName><forename type="first">L</forename><surname>Despalatović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Vojković</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vukicevic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2014 37th international convention on information and communication technology, electronics and microelectronics</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="997" to="1002" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Ictnet at trec 2019 news track</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Lian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Hou</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Fast and accurate annotation of short texts with wikipedia pages</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ferragina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Scaiella</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE software</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="70" to="75" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Smith at trec2019: Learning to rank background articles with poetry categories and keyphrase extraction</title>
		<author>
			<persName><forename type="first">J</forename><surname>Foley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montoly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pena</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A set of measures of centrality based on betweenness</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Freeman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Sociometry</title>
		<imprint>
			<biblScope unit="page" from="35" to="41" />
			<date type="published" when="1977">1977</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Multiple additive regression trees with application in epidemiology</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Meulman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistics in medicine</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="1365" to="1381" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Community structure in social and biological networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Girvan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Newman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the national academy of sciences</title>
		<imprint>
			<biblScope unit="volume">99</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="7821" to="7826" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Improving ad hoc retrieval with bag of entities</title>
		<author>
			<persName><forename type="first">G</forename><surname>Gonçalves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Magalhães</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Callan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">image</title>
		<imprint>
			<biblScope unit="volume">409</biblScope>
			<biblScope unit="issue">68</biblScope>
			<biblScope unit="page">116</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Exploring network structure, dynamics, and function using networkx</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hagberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Swart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chult</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<pubPlace>Los Alamos, NM (United States</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Los Alamos National Lab. ; LANL)</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Tech. rep.</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">On the reproducibility of the tagme entity linking system</title>
		<author>
			<persName><forename type="first">F</forename><surname>Hasibi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Balog</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Bratsberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="436" to="449" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Trec 2018 news track</title>
		<author>
			<persName><forename type="first">S</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Soboroff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NewsIR@ ECIR</title>
				<imprint>
			<date type="published" when="2018">2079. 2018</date>
			<biblScope unit="page" from="57" to="59" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Paragraph as lead-finding background documents for news articles</title>
		<author>
			<persName><forename type="first">K</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Fang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Leveraging entities in background document retrieval for news articles</title>
		<author>
			<persName><forename type="first">K</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Fang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Dminr at trec news track</title>
		<author>
			<persName><forename type="first">S</forename><surname>Missaoui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Macfarlane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Makri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gutierrez-Lopez</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Relevance score normalization for metasearch</title>
		<author>
			<persName><forename type="first">M</forename><surname>Montague</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Aslam</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the tenth international conference on Information and knowledge management</title>
				<meeting>the tenth international conference on Information and knowledge management</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="427" to="433" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Unc sils at trec 2019 news track</title>
		<author>
			<persName><forename type="first">J</forename><surname>Qu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>TREC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">The probabilistic relevance framework: BM25 and beyond</title>
		<author>
			<persName><forename type="first">S</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zaragoza</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>Now Publishers Inc</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Soboroff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</author>
		<title level="m">Trec 2018 news track overview</title>
				<imprint>
			<publisher>TREC</publisher>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">N</forename><surname>Milne</surname></persName>
		</author>
		<title level="m">An effective, low-cost measure of semantic relatedness obtained from wikipedia links</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Bag-of-entities representation for ranking</title>
		<author>
			<persName><forename type="first">C</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Callan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">Y</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 ACM International Conference on the Theory of Information Retrieval</title>
				<meeting>the 2016 ACM International Conference on the Theory of Information Retrieval</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="181" to="184" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Word-entity duet representations for document ranking</title>
		<author>
			<persName><forename type="first">C</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Callan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">Y</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th International ACM SIGIR conference on research and development in information retrieval</title>
				<meeting>the 40th International ACM SIGIR conference on research and development in information retrieval</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="763" to="772" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Anserini at trec 2018: Centre, common core, and news tracks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Seventh Text REtrieval Conference (TREC 2018)</title>
				<meeting>the Twenty-Seventh Text REtrieval Conference (TREC 2018)<address><addrLine>Gaithersburg, MD</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
