<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">An Interactive Two-Dimensional Approach to Query Aspects Rewriting in Systematic Reviews. IMS Unipd At CLEF eHealth Task 2</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Giorgio</forename><forename type="middle">Maria</forename><surname>Di</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Information Engineering</orgName>
								<orgName type="institution">University of Padua</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Federica</forename><surname>Beghini</surname></persName>
							<email>fede.beghini92@gmail.com</email>
							<affiliation key="aff1">
								<orgName type="department">Dept. of Linguistic and Literary Studies</orgName>
								<orgName type="institution">University of Padua</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Federica</forename><surname>Vezzani</surname></persName>
							<email>federicavezzani92@gmail.com</email>
							<affiliation key="aff1">
								<orgName type="department">Dept. of Linguistic and Literary Studies</orgName>
								<orgName type="institution">University of Padua</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Geneviève</forename><surname>Henrot</surname></persName>
							<email>genevieve.henrot@unipd.it</email>
							<affiliation key="aff1">
								<orgName type="department">Dept. of Linguistic and Literary Studies</orgName>
								<orgName type="institution">University of Padua</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">An Interactive Two-Dimensional Approach to Query Aspects Rewriting in Systematic Reviews. IMS Unipd At CLEF eHealth Task 2</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E756195E274FBC9B99F2371A177AE51C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we describe the participation of the Information Management Systems (IMS) group at CLEF eHealth 2017 Task 2. This task focuses on the problem of systematic reviews, that is articles that summarise all evidence that is published regarding a certain medical topic. This task, known in Information Retrieval as the total recall problem, requires long and tedious search sessions by experts in the field of medicine. Automatic (or semi-automatic) approaches are essential to support these type of searches when the amount of data exceed the limits of users, i.e. in terms of attention or patience. We present the two-dimensional probabilistic version of BM25 with explicit relevance feedback together with a query aspect rewriting approach for both the simple evaluation and the cost-effective evaluation.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In this paper, we describe the participation of the Information Management Systems (IMS) group at CLEF eHealth 2017 <ref type="bibr" target="#b0">[1]</ref> Task <ref type="bibr" target="#b1">[2]</ref>. This task focuses on the problem of systematic reviews, that is articles that summarise all evidence that is published regarding a certain medical topic. This task, known in Information Retrieval as the total recall problem, requires long and tedious search sessions by experts in the field of medicine. Automatic (or semi-automatic) approaches are essential to support these type of searches when the amount of data exceed the limits of users, i.e. in terms of attention or patience. In particular, the aim is to make research papers abstract and title screening more effective given the results of a boolean search submitted to a medical database.</p><p>The CLEF eHealth Task 2 has two types of evaluation procedures to assess the quality of a system that supports systematic reviews. These procedures are based on the following assumptions:</p><p>-Simple evaluation, the user of the system is the researcher (end-user) that performs the abstract and title screening of the retrieved articles. Every time the system returns an abstract to the end-user there is an incurred cost.</p><p>-Cost-effective evaluation, the user that performs the screening is not the end-user. The user can interchangeably perform abstract and title screening, or document screening, and decide what documents to pass to the end-user. Every time the system provides an abstract to the user, she/he can i) either read the abstract (with an incurred cost, like in the simple evaluation) and decide whether to pass this document to the end-user, ii) or read the full document (with a higher cost) and decide whether to pass this document to the end-user, iii) or directly pass the document to the end-user. For each document passed to the end-user there are additional costs that need to be added.</p><p>The objective of our participation to this task was to:</p><p>find the best parameters (in terms of classification/ranking accuracy) of the BM25 model <ref type="bibr" target="#b3">[4]</ref>; explore the problem of query aspects and query (re-)formulation given an information need <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b9">10]</ref>; integrate the query aspects into the two-dimensional probabilistic model <ref type="bibr" target="#b2">[3]</ref>; study an automatic feedback loop to find the optimal stopping strategy <ref type="bibr" target="#b7">[8]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Approach</head><p>In this paper, we continue to investigate the interaction with the two dimensional interpretation of the BM25 model applied to the problem of explicit relevance feedback with three goals in mind:</p><p>we want to create a set of relevance judgements with the least effort by human assessors, we use interactive visualizations to interpret the intermediate results of the relevance feedback, we use explicit query rewriting by non experts to create different aspects of the information need.</p><p>Following the work started in <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b6">7]</ref>, we continue to study the two-dimensional interpretation of the BM25 model defined in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">BM25</head><p>The BM25 is a probabilistic retrieval model where, if we use the definition given by Zaragoza and Robertson in <ref type="bibr" target="#b8">[9]</ref>, the weight of the i-th term in a document is equal to:</p><formula xml:id="formula_0">w BM 25 i (tf ) = tf k 1 (1 − b) + b dl avdl + tf w BIM i (1)</formula><p>where k 1 and b are two parameters (we used the default values used by Terrier 3 , k 1 = 1.2 and b = 0.75), tf is the term frequency in the document, and w BIM i is 3 http://terrier.org the Binary Independence Model weight of the i-th term:</p><formula xml:id="formula_1">w BIM i = log θ R i (1 − θ R i ) (1 − θ N R i ) θ N R i (2)</formula><p>where θ R i and θ N R i are the parameters of the Bernoulli random variable that represent the presence (or absence) of the i-th term in the relevant (R) and non-relevant (N R) documents. The estimate of each parameter is:</p><formula xml:id="formula_2">θ R i = r i + α R R + α R + β R<label>(3)</label></formula><formula xml:id="formula_3">θ N R i = n i − r i + α N R N − R + α N R + β N R (<label>4</label></formula><formula xml:id="formula_4">)</formula><p>where R is the number of relevant documents, r i the number of relevant documents in which the i-th term appears, N is the total number of documents and n i is the total number of documents in which the i-th term appears. Parameters α and β correspond to the hyper-parameter of the conjugate beta prior distribution of the Bernoulli random variable. For α R = β R = 0.5 and β R = N R = 0.5, we obtain the definition of the well-known Robertson -Spärck Jones weight</p><formula xml:id="formula_5">w RSJ i [9].</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Two-Dimensional Model</head><p>The two-dimensional representation of probabilities <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b7">8]</ref> is an intuitive way of presenting a two-class classification problem on a two-dimensional space. Given two classes, for example relvant R and non-relevant N R, a document d is assigned to category R if the following inequality holds:</p><formula xml:id="formula_6">P (d|N R) y &lt; m P (d|R) x +q<label>(5)</label></formula><p>where P (d|R) and P (d|N R) are the likelihoods of the object d given the two categories, while m and q are two parameters that can be assigned (automatically or by a user) to compensate for either the unbalanced class issues or different misclassification costs.</p><p>If we interpret the two likelihoods as two coordinates x and y of a two dimensional space, the problem of classification can be studied on a two-dimensional plot. The decision of the classification is represented by the line y = mx + q that splits the plane into two parts: all the points that fall 'below' this line are classified as objects that belong to class R.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Two-dimensional BM25</head><p>In order to link the two-dimensional model to the BM25 model, first we define the BIM weight as a difference of logarithms:</p><formula xml:id="formula_7">w BIM i = log θ R i (1 − θ R i ) − log θ N R i (1 − θ N R i ) = w BIM,R i − w BIM,N R i (6)</formula><p>then, we can define the BM25 term weight accordingly</p><formula xml:id="formula_8">w BM 25 i (tf ) = tf k 1 (1 − b) + b dl avdl + tf w BIM,R i − w BIM,N R i<label>(7)</label></formula><p>We now have all the elements to define the two coordinates x = P (d|R) and y = P (d|N R) in the following way:</p><formula xml:id="formula_9">P (d|R) = i∈d w BM 25,R i (tf )<label>(8)</label></formula><formula xml:id="formula_10">P (d|N R) = i∈d w BM 25,N R i (tf )<label>(9)</label></formula><p>where i∈d indicates (with an abuse of notation) the sum over all the terms of document d.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Method</head><p>Given the definition of two-dimensional BM25 model, we focused on the following problems:</p><p>1. find the best combination of hyper-parameters α R , α N R , β R , β N R , 2. devise a strategy to create different query aspects of the same information need given a minimum amount of relevance feedback, 3. explore different options of explicit relevance feedback for both the simple and the cost-effective evaluation subtasks.</p><p>For step 1), we used the training data available for this task to find the best combination of parameters trough a force brute approach.</p><p>For step 2), we decided to use the following procedure:</p><p>for each topic, we run a plain BM25 retrieval model and get the relevance feedback for the first abstract in the ranking list, we get the explicit relevance feedback on that abstract and ask to two different people (non-experts in the field of medicine) to review the abstract and rewrite an alternative query also according to the value of the feedback (relevant or not), For example, for topic CD008803 the original information need is expressed by the following statement:</p><p>"Optic nerve head and fibre layer imaging for diagnosing glaucoma"</p><p>we run BM25 and obtain the top retrieved abstract is document 19028735, the content of which is:</p><p>title: Imaging of the retinal nerve fibre layer for glaucoma. abstract: Glaucoma is a group of diseases characterised by retinal ganglion cell dysfunction and death. Detection of glaucoma and its progression are based on identification of abnormalities or changes in the optic nerve head (ONH) or the retinal nerve fibre layer (RNFL), either functional or structural. This review will focus on the identification of structural abnormalities in the RNFL associated with glaucoma. . . .</p><p>Then we pass the information that this abstract is not relevant (according to the abstract qrels) to the two users that rewrite the query, and we obtain two new query aspects. One user was asked to prepare a list of terms:</p><p>"optic nerve head, ONH, optic disc, fibre layer, diagnosis, retinal, imaging, RNFL, glaucoma, SLP, Scanning laser polarimetry, HRT, Heidelberg Retina Tomograph, OCT, Optical Coherence Tomography, GDx"</p><p>The other user had to write a sort of information need instead of a list of words:</p><p>"Diagnostic accuracy of HRT, OCT and GDx for diagnosing manifest glaucoma by detecting ONH and RNFL damage."</p><p>The first type of query was written with the aim of entering the key words contained in the topic title, in the boolean query and in the article that was given (if relevant), along with other terms which were the result of various processes: the componential analysis of some words, the variants, the synonyms, the declensions and the acronyms of some terms and the reading of other relevant information using sources on the web <ref type="foot" target="#foot_0">4</ref> . The componential analysis consists of breaking down the sememe (i.e. the meaning) of a word in all its sense components <ref type="foot" target="#foot_1">5</ref> , e.g. the semes of radiculopathy<ref type="foot" target="#foot_2">6</ref> (topic CD007431) are /pathology/, /nerve root/, /spinal/, /inflammation/, /compression/. Therefore, in this case we also included all these terms in the query, which were not present in the information need <ref type="foot" target="#foot_3">7</ref> . We did not decide to enter the semes of all the words, but just of the terms whose semes we thought could improve the search of the most relevant articles. Furthermore, if the terms had many variants, we added them to the query: e.g. in topic CD008760, we did not just enter oesophageal and oesophagus, but also esophageal and esophagus. Moreover, we tried to use not only one grammatical form to describe a concept, which is why we did not just enter nouns, but also verbs and adjectives, e.g. radiculopathy, radicular and spinal, spine (topic CD007431); endometriosis, endometrial (topic CD012019), diagnosis, diagnose, diagnosing, diagnostic (topic CD010542). We also added synonyms, e.g. diagnosis, screening, examination (topic CD009925) and diagnosis, detection CD010783). For what concerns the process of declension, sometimes we wrote not only the singular, but also the plural form of a noun, e.g. dementia, dementias; biomarker, biomarkers (topic CD008782). Then, we entered the acronym of some terms, e.g. LPB (lumbago) (topic CD007431); mild cognitive impairment (MCI) (CD008782). Finally, the terms have been entered in a random order.</p><p>The second type of query was written with the aim of reformulating the information need. The purpose was to rewrite the information given for each topic using an alternative terminology and trying to reformulate a meaningful and humanly readable sentence. For this reason, validly attested synonyms and orthographic alternatives were used as variants of the medical terms provided in the original information need. In addition, another criterion was to systematically replace acronyms with their expansions and expansions with their acronyms. For example, for topic CD009135, the information need "Rapid tests for the diagnosis of visceral leishmaniasis in patients with suspected disease" was rewritten using synonyms and acronyms for "visceral leishmaniasis": "Evaluation of rapid examinations in order to detect VL (kala-azar, black fever and Dumdum fever) in patients with clinically suspected infection". This approach allowed us to expand the medical terminology and to evaluate also the documents in which the selected variants were present. The sources from which the terminological variants were selected were PubMed, the online medical dictionary Merriam Webster and Wikipedia. For what concerns the topics presenting a relevant document (relevance index 1) selected by the expert, the criterion of re-writing the information need was based on the knowledge acquired by reading the PubMed article abstract. This step facilitated the reformulation of the title by focusing on the typology of the request and its related aspects. On the contrary, the topics where the document's relevance index was 0, the reformulation was based on the terminology frequency analysis and on an in-depth research of the topic on reliable sources available on the web.</p><p>For step 3), we designed alternative strategies that use the following parameters:</p><p>number of documents to assess, in batches or iteratively, percent of documents to assess, maximum number of documents to assess per iteration, number of terms to add at each feedback iteration, for the cost-effective evaluation, the minimum precision the system can reach before stopping the search.</p><p>Simple evaluation For the simple evaluation subtask, we focused on the number (or percentage) of documents to use for explicit relevance feedback and how to combine the query aspects. No threshold on the number of documents to retrieve was set.</p><p>Cost-effective evaluation For the cost-effective subtask, we performed two rounds of relevance feedback: first retrieve, then classify. In the first round, we select a percentage of documents for explicit relevance feedback; then, we use the relevance information to build the two classes R and N R. Once the two classes are built, we use the two-dimensional space to pick the document with partial recall 100% (by 'partial', mean that if during the iteration we retrieve 10 relevant document out of 20, we pick the relevant document with the lowest score) and let the classification line pass through that point. Then we iterate the feedback until precision reaches 0.2.</p><p>In Figure <ref type="figure">1</ref>, we show the two dimensional situation at four different steps of the iteration. Green dots represents relevant documents, red dots non-relevant documents, black dots documents to be ranked (or judged). In Figure <ref type="figure">1 (a)</ref>, we see the documents at the end of the relevance feedback phase. After we re-set the probabilities by building the two classes of relevant and non relevant documents, the documents are in a different position of the two-dimensional space, Figure <ref type="figure">1</ref> (b). The space between the interpolating line of the relevant documents (dashed line) and the line of the last relevant document (dot-dashed line) is the 'grey area' where we expect to find more relevant documents. After a few iteration, the relevant and non relevant clouds of points become more and more separate, Figure <ref type="figure">1 (c)</ref>. When all the documents within the space between the two lines are judged (plus some other of the 'extra-rounds') the systems stops sending documents to the user, Figure <ref type="figure">1 (d)</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experiments</head><p>In all experiments, we used the first document retrieved with a BM25 approach (and then judged) to build two different queries that represent the same information need. The two alternative queries are combined with the original one in different ways as described in the following sections.</p><p>For all the experiments, we set the best set of values for the parameters α R , α N R , β R , β N R of the BM25 found with a brute force approach on the training data. The values are:</p><formula xml:id="formula_11">-α R = α N R = 1.0 -β R = β N R = 0.01</formula><p>These values are consistent with other experiments and indicate that a beta prior distribution that discounts the 'presence' of a term in favour of its 'absence' (high α and low β) results in a better retrieval performance.</p><p>We also run a set of experiments on the training data to find the value of the number of documents k to use for relevance feedback that gives the best Fig. <ref type="figure">1</ref>: Cost-effective approach on the two-dimensional space. Green dots represent relevant documents, red dots non-relevant documents, black dots documents to be ranked (or judged). The dashed line shows the interpolating line of the relevant documents, while the dot-dashed line indicates the last relevant documents found. When all the documents within this space are judged (plus some other of the 'extra-rounds') the systems stops sending documents to the user.</p><p>trade-off between cost and effectiveness, and we found that k = 50 is a good estimate.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Simple Evaluation</head><p>For the simple evaluation subtask, we submitted four runs:</p><p>-ims iafa m10k150f0m10, run-1, this run uses Interactive Automatic Feedback with query Aspects (iafa) and, for each topic, uses k = 150 feedback rounds where, at each round, a new word is picked from the relevant documents and the top document is judged. For each topic, a total of 150 documents are judged. -ims iafas m10k50f0m10, run-2, this run uses Interactive Automatic Feedback with query Aspects with Separate rankings (iafas). At each round of feedback, the two query variants are run in parallel with the original one and three different documents are judged. There are k = 50 rounds for a total of 150 documents judged per topic. -ims iafap m10p2f0m10, run-3, this run uses Interactive Automatic Feedback with query Aspects using a Percent (iafap) of documents for feedback. This run is similar to the first one but it uses a number of documents for relevance feedback that is proportional to the number of documents to rank. In this case, p = 2 is two percent of feedback. -ims iafap m10p5f0m10, run-4, this run uses Interactive Automatic Feedback with query Aspects using a Percent (iafap) of documents for feedback. The percent of feedback is p = 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Cost-Effective Evaluation</head><p>For the cost-effective evaluation subtask, we submitted four runs. All the four runs use the same approach named Interactive Automatic Feedback with query Aspects and Percent of relevance feedback and Classification (iafapc). In particular, we tried different values of parameters concerning the percent of documents for relevance feedback and the maximum number of documents for relevance feedback in the initial phase.</p><p>During the classification phase, we calculate the linear interpolation of relevant documents if 5 or more relevant documents are available, otherwise we compute the linear interpolation of non relevant documents. If the angular coefficient of the line is less than 0.9, we adjust it. We iterate this process by selecting the top 10 documents and perform explicit relevance feedback until precision reaches 0.2. After that point, extra iterations are performed with half of the documents used in the previous feedback round. We stop if no other documents are available or precision is below 0.2 and we have only one document for the extra rounds of relevance feecback.</p><p>-ims iafapc m10p5f0t0p2m10, run-5, this run uses 5 percent of relevance feedback documents per round of relevance feedback in the initial phase. -ims iafapc m10p10f0t150p2m10, run-6, this run uses 10 percent of relevance feedback and a maximum of 150 documents per round of relevance feedback in the initial phase. -ims iafapc m10p20f0t150p2m10, run-7, this run uses 20 percent of relevance feedback and a maximum of 150 documents per round of relevance feedback in the initial phase. -ims iafapc m10p20f0t300p2m10, run-8, this run uses 20 percent of relevance feedback and a maximum of 300 documents per round of relevance feedback in the initial phase.</p><p>The results for the simple evaluation are reported in Table <ref type="table" target="#tab_0">1</ref> and Figure <ref type="figure" target="#fig_1">2a</ref> while the results for the cost-effective evaluation are reported in Table <ref type="table" target="#tab_1">2</ref> and Figure <ref type="figure" target="#fig_1">2b</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Final Remarks and Future Work</head><p>In this paper, we presented the experiments of our research group to the CLEF eHealth Task 2. The objective of our participation to this task was to investi- gate a better set of parameters for the BM25, explore the problem of query aspects and query (re-)formulation given an information need, integrate the query aspects into the two-dimensional probabilistic model, and study an automatic feedback loop to find the optimal stopping strategy. Some interesting findings during the training phase that we will document more deeply in the future can be summarised as follows:</p><p>there are values for the α and β parameter that clearly outperform the standard BM25 with α = β = 0.5; performing an iterative explicit relevance feedback one document at a time changes significantly the performance of both retrieval and classification (the cost of training at each round of feedback is very high in computational terms, though); adding query aspects to the original information need increase consistently the performance of both the retrieval and classification; choosing the right terms to add during the iteration of relevance feedback may change significantly the results of both the retrieval and classification.</p><p>The results of the test phase presented in the previous section will be analyzed more deeply in the next weeks. In particular, it is not clear whether a fixed amount of documents (k = 150, for example) may be better than a fixed percentage of documents (say, p = 5). It will be interesting to study and compare the simple and the cost-effective strategies in terms of the actual costs, as shown by Table <ref type="table" target="#tab_1">1 and Table 2</ref>. We will also continue to investigate the process of query aspect rewriting by extending it to the case of iteratively rewriting the query aspects according to the shifts of the two-dimensional cloud of points, as shown in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 :</head><label>2</label><figDesc>Fig. 2: NCG at different recall values for the simple and cost-effective evaluation.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Simple evaluation results. Top part shows abstract qrels evaluation, bottom part document qrels evaluation.</figDesc><table><row><cell>run</cell><cell cols="3">ap last rel wss100 wss95 norm area total cost loss er loss r</cell></row><row><cell cols="2">run-1 .280 2269.333</cell><cell>.415 .508</cell><cell>.896 4075.233 .544 .000</cell></row><row><cell cols="2">run-2 .266 2304.600</cell><cell>.410 .517</cell><cell>.892 4206.567 .544 .000</cell></row><row><cell cols="2">run-3 .253 2395.533</cell><cell>.366 .476</cell><cell>.875 4076.367 .544 .000</cell></row><row><cell cols="2">run-4 .269 2260.467</cell><cell>.398 .496</cell><cell>.885 4311.433 .544 .000</cell></row><row><cell cols="2">run-1 .223 1055.793</cell><cell>.706 .713</cell><cell>.932 3935.414 .544 .000</cell></row><row><cell cols="2">run-2 .190 990.000</cell><cell>.706 .723</cell><cell>.928 4065.345 .544 .000</cell></row><row><cell cols="2">run-3 .202 838.897</cell><cell>.661 .685</cell><cell>.919 4156.517 .544 .000</cell></row><row><cell cols="2">run-4 .212 1007.379</cell><cell>.706 .703</cell><cell>.931 4311.433 .544 .000</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Cost-effective evaluation results. Top part shows abstract qrels evaluation, bottom part document qrels evaluation.</figDesc><table><row><cell>run</cell><cell cols="4">ap last rel wss100 wss95 norm area total cost total uni total wei loss er loss r</cell></row><row><cell cols="2">run-5 .232 540.400</cell><cell>.151 .176</cell><cell cols="2">.755 653.033 1488.064 5021.993 .115 .097</cell></row><row><cell cols="2">run-6 .244 379.533</cell><cell>.133 .168</cell><cell cols="2">.774 478.300 2228.266 6008.652 .124 .053</cell></row><row><cell cols="2">run-7 .264 396.800</cell><cell>.183 .247</cell><cell cols="2">.808 511.633 2238.965 5969.683 .161 .044</cell></row><row><cell cols="2">run-8 .270 615.933</cell><cell>.255 .411</cell><cell cols="2">.859 807.300 1714.993 4813.495 .169 .019</cell></row><row><cell cols="2">run-5 .190 364.034</cell><cell>.428 .529</cell><cell cols="2">.845 633.931 1029.042 2292.358 .091 .063</cell></row><row><cell cols="2">run-6 .200 280.345</cell><cell>.385 .441</cell><cell cols="2">.856 474.966 1365.765 3813.256 .119 .023</cell></row><row><cell cols="2">run-7 .216 280.379</cell><cell>.480 .535</cell><cell cols="2">.886 509.448 1388.187 3833.256 .170 .017</cell></row><row><cell cols="2">run-8 .217 414.586</cell><cell>.578 .638</cell><cell>.913</cell><cell>793.69 1310.866 2615.199 .206 .007</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">PubMed https://www.ncbi.nlm.nih.gov/pubmed/ The Free Dictionary by Farlex -Medical Dictionary http://medical-dictionary. thefreedictionary.com/radiculopathy Merriam Webster Dictionary https://www.merriam-webster.com/) Wikipedia https://en.wikipedia.org/wiki/Main_Page</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">Rastier, F, (1987), Smantique interprtative, d. Presses Universitaires de France, 2009, Paris, p.18-32. Dubois., J. et al. (1994), Dictionnaire de linguistique et des sciences du langage, d.Larousse, Paris, p.423-424. Ducrot, O., Schaeffer, J.-M., (1972), Nouveau dictionnaire encyclopdique des sciences du langage, d.du Seuil, 1995, Paris, p.445-447.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">The Free Dictionary by Farlex -Medical Dictionary http://medical-dictionary. thefreedictionary.com/radiculopathy</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_3">Physical examination for lumbar radiculopathy due to disc herniation in patients with low-back pain.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">CLEF 2017 eHealth Evaluation Lab Overview</title>
		<author>
			<persName><forename type="first">Lorraine</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Liadh</forename><surname>Kelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hanna</forename><surname>Suominen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aurélie</forename><surname>Névéol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aude</forename><surname>Robert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Evangelos</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rene</forename><surname>Spijker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">João</forename><surname>Palotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guido</forename><surname>Zuccon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF 2017 -8th Conference and Labs of the Evaluation Forum</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">CLEF 2017 Technologically Assisted Reviews in Empirical Medicine Overview</title>
		<author>
			<persName><forename type="first">Evangelos</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dan</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leif</forename><surname>Azzopardi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rene</forename><surname>Spijker</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2017 -Conference and Labs of the Evaluation forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">September 11-14, 2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A new decision to take for cost-sensitive naïve bayes classifiers</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Inf. Process. Manage</title>
		<imprint>
			<biblScope unit="volume">50</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="653" to="674" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Geometric perspectives of the BM25</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th Italian Information Retrieval Workshop</title>
				<meeting>the 6th Italian Information Retrieval Workshop<address><addrLine>Cagliari, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">May 25-26, 2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Interactive text categorisation: The geometry of likelihood spaces</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Studies in Computational Intelligence</title>
		<imprint>
			<biblScope unit="volume">668</biblScope>
			<biblScope unit="page" from="13" to="34" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Gamification for IR: the query aspects game</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maria</forename><surname>Maistro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Zilio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Third Italian Conference on Computational Linguistics (CLiC-it 2016) &amp; Fifth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2016)</title>
				<meeting>Third Italian Conference on Computational Linguistics (CLiC-it 2016) &amp; Fifth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA 2016)<address><addrLine>Napoli, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">December 5-7, 2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Gamification for machine learning: The classification game</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maria</forename><surname>Maistro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Zilio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third International Workshop on Gamification for Information Retrieval co-located with 39th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2016)</title>
				<meeting>the Third International Workshop on Gamification for Information Retrieval co-located with 39th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2016)<address><addrLine>Pisa, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-07-21">July 21, 2016. 2016</date>
			<biblScope unit="page" from="45" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The university of padua (IMS) at TREC 2016 total recall track</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Di</forename><surname>Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maria</forename><surname>Maistro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Zilio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Twenty-Fifth Text REtrieval Conference, TREC 2016</title>
				<meeting>The Twenty-Fifth Text REtrieval Conference, TREC 2016<address><addrLine>Gaithersburg, Maryland, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">November 15-18, 2016, 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The probabilistic relevance framework: BM25 and beyond</title>
		<author>
			<persName><forename type="first">Stephen</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hugo</forename><surname>Zaragoza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and Trends in Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="333" to="389" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Scentbar: A query suggestion interface visualizing the amount of missed relevant information for intrinsically diverse search</title>
		<author>
			<persName><forename type="first">Kazutoshi</forename><surname>Umemoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Takehiro</forename><surname>Yamamoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Katsumi</forename><surname>Tanaka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2016</title>
				<meeting>the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2016<address><addrLine>Pisa, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">July 17-21, 2016. 2016</date>
			<biblScope unit="page" from="405" to="414" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
