<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Technology-Assisted Review in Empirical Medicine: Waterloo Participation in CLEF eHealth 2017</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gordon</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Cheriton School of Computer Science</orgName>
								<orgName type="institution">University of Waterloo</orgName>
								<address>
									<postCode>N2L 3G1</postCode>
									<settlement>Waterloo</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maura</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Cheriton School of Computer Science</orgName>
								<orgName type="institution">University of Waterloo</orgName>
								<address>
									<postCode>N2L 3G1</postCode>
									<settlement>Waterloo</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Technology-Assisted Review in Empirical Medicine: Waterloo Participation in CLEF eHealth 2017</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">C85B7BDBC7985786233485A42C5C98BA</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Screening articles for studies to include in systematic reviews is an application of technology-assisted review ("TAR"). In this work, we applied the Baseline Model Implementation ("BMI") from the TREC  Total Recall Track (2015-2016)  to the CLEF eHealth 2017 task of screening MEDLINE abstracts to identify articles reporting studies to be considered for inclusion. According to rank-based evaluation measures, this approach identified every article describing a study that should have been included in each of 30 systematic reviews, by examining 461 abstracts, on average, per review-12.6% of the 3, 655 abstracts that would have had to be examined, on average, if instead, a manual approach had been used. While this result indicates TAR's promise to substantially reduce the time and cost of abstract screening, this promise can be realized only if it can be known with reasonable certainty for each review how many abstracts must be examined before all, or substantially all, articles that should be included have been identified. To this end, we applied our "knee-method" stopping criterion to BMI to determine how many abstracts should be examined for each topic. According to thresholdbased evaluation, the knee method identified every article that should have been included (100% recall), while examining 2, 659 abstracts, on average, per topic-72.8% of the 3, 655 abstracts, that would have required examination, on average, had a manual approach been used instead. While our results suggest that TAR can substantially improve the efficiency of abstract screening without compromising recall, there remains room for improvement both in ranking and stopping criterion, as well as important factors that were not addressed in the CLEF eHealth 2017 framework: the completeness of the universe of abstracts gathered using keyword search, and the accuracy of the human assessments of the collected abstracts.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The University of Waterloo participated in Task 2, Technologically Assisted Reviews in Empirical Medicine <ref type="bibr" target="#b9">[10]</ref>, of the CLEF 2017 eHealth Evaluation Lab <ref type="bibr" target="#b11">[12]</ref>. Task 2 simulates the second phase-screening-in a prototypical three-phase workflow to identify studies for inclusion in a systematic review:</p><p>1. Search: First, Boolean queries are used to identify as many articles as possible that may describe studies that should be included; 2. Screening: Second, titles and abstracts of the articles identified in the search phase are examined to eliminate those which could not possibly describe studies that should be included; and 3. Selection: Finally, articles that survived the screening phase are read in full to determine whether or not they meet the systematic review inclusion criteria.</p><p>The overall objective of our research is to improve the human efficiency, as well as the effectiveness, of workflows to identify studies for inclusion in systematic reviews. The results of our CLEF experiments support the hypothesis that continuous active learning ("CAL") can substantially improve the human efficiency of screening, without substantially compromising its effectiveness. The results also are consistent with the further hypothesis that CAL actually improves effectiveness by identifying articles missed in the search phase, or articles mistakenly eliminated during the screening phase. While this hypothesis cannot be tested immediately within the framework of Task 2, we have identified a set of articles that, were it determined that they describe one or more studies that should have been included in the review, would demonstrate CAL's superior effectiveness. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Apparatus</head><p>Task 2 is essentially the Technology-Assisted Review ("TAR") task addressed by the TREC 2015 and TREC 2016 Total Recall Tracks <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b7">8]</ref>. For our participation in CLEF, we reprised our Total Recall efforts using the same apparatus. At TREC, the systems under test were given, at the outset, a corpus of documents and a set of topics. For each topic, a system under test repeatedly submitted documents from the corpus to a server, and in return, was given a simulated human assessment of "relevant" or "not relevant" for each document.</p><p>The objective was to identify as many relevant documents as possible, while submitting as few non-relevant documents as possible. The tension between these Algorithm 1 The AutoTAR Continuous Active Learning ("CAL") Method, as Implemented by the TREC Baseline Model Implementation ("BMI") and deployed by Waterloo for the CLEF Technologically Assisted Review Task. two criteria was evaluated using rank-based measures (e.g., recall as a function of the number of documents submitted), as well as set-based measures (e.g., recall at a point when a certain number of documents, specified contemporaneously by the system, had been submitted).</p><p>Prior to TREC, we made available a Baseline Model Implementation ("BMI"), 1 to illustrate the client-server protocol, as well as to provide baseline results for comparison. BMI, which encapsulates our AutoTAR Continuous Active Learning ("CAL") method <ref type="bibr" target="#b0">[1]</ref>, yielded rank-based results that compared favorably will all systems under test. During the course of our participation in TREC, we developed and tested the "knee method" stopping procedure <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b4">5]</ref>, with the purpose of achieving high recall with high probability.</p><p>Task 2 differed operationally from the TREC Total Recall Track in that a list of document identifiers, rather than a corpus, was supplied at the outset, and a complete set of relevance assessments, rather than an assessment server were used to simulate human assessments. Task 2 also differed substantively from the Total Recall Track in that the corpus for each topic was narrowed by a search phase specific to that topic, and therefore yielded a much smaller set that was richer in relevant documents. Task 2 differed further in that two sets of relevance assessments were available: the assessments from a previously conducted screening phase, and the assessments from a previously conducted selection phase, raising the question of which assessments (or combination of assessments) should be used to simulate relevance feedback, and which should be used to evaluate the results (cf. <ref type="bibr" target="#b5">[6]</ref>).</p><p>Task 2 provides no method equivalent to TREC's "call your shot" for a system under test to specify a stopping criterion (for threshold-based evaluation), while at the same time continuing until every document in the corpus has been submitted for assessment (for rank-based evaluation).</p><p>Task 2, however, unlike TREC, afforded participants the opportunity to conduct task-specific tuning and configuration, by supplying 20 training topics (with corresponding corpora and assessments) in advance of the exercise, followed by 30 test topics, which were used for evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Training and Configuration</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Document Corpora</head><p>The corpus for each topic consisted of abstracts from MEDLINE/Pubmed 2 identified by PMID. On March 8, 2017, we fetched the entire MEDLINE dataset consisting of 27,348,935 XML files, each containing the titles, abstracts, and metadata for an article. We used the raw XML files as documents in the corpora that were supplied at the outset to BMI. Our original intent had been to apply BMI to the entire corpus of 27,348,935 files, thus combining the search and screening phases. When we employed this strategy in a pilot experiment on the test topics, we found that no assessments were available for many, if not most, of the highly ranked documents returned by BMI. To our eye, these documents were indistinguishable from those for which "relevant" assessments were provided. We investigated, without success, the reasons why these documents were not retrieved by the previously conducted search phase. For example, the documents in question were neither newer nor older than those for which assessments were available, and appeared to contain relevant terms from the search query. As we were unable to reproduce the results of the CLEF search phase, we chose to ignore-for the purpose of relevance feedback and evaluation-documents for which no assessments were available. Ignoring these unjudged documents, our pilot experiment yielded what appeared to be reasonable rank-based results.</p><p>Ignoring documents for feedback and evaluation yields a substantially different result from removing them from the corpus altogether. In a second pilot experiment, we constructed a separate corpus for each topic, consisting of only those documents for which relevance assessments were available. While BMI ran much faster on these reduced corpora than on the 27M dataset, results were apparently inferior. We conjecture that this inferior result can be explained by skewed term-frequency statistics in the reduced corpora.</p><p>As a compromise between the effectiveness of searching the 27M dataset and the (computational) efficiency of searching the reduced corpora, we conducted a third pilot experiment using a common corpus consisting of all documents that were assessed for any of the 20 test topics. That is, for any given topic, the corpus consisted of all the documents assessed for that topic, as well as all the documents assessed for each of the other 19 topics. Our rationale was that including documents retrieved for all topics would introduce enough diversity to unskew sufficiently the term-frequency statics. This approach appeared to achieve the efficiency of using reduced corpora and the effectiveness of using the full dataset, and was chosen for our official tests: For the official tests, the corpus consisted of all documents assessed for any of the 30 test topics (less four documents whose PMIDs were not present in our MEDLINE database); from this corpus, we submitted and solicited feedback only for documents for which assessments were available.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Relevance Feedback</head><p>We investigated three modes of relevance feedback, of which only two were selected for official testing:</p><p>1. Relevance feedback based on the screening-phase assessments (selected as Method A for official testing); 2. Relevance feedback based on the selection-phase assessments (not selected for official testing); 3. Relevance feedback based on a hybrid of screening-phase and selection-phase assessments (selected as Method B for official testing).</p><p>The first and second methods are straightforward: When BMI identifies a document for assessment, the judgment returned to BMI is that supplied by CLEF for either the screening phase (the "abstract qrels") or the selection phase ("the content qrels"). The third method operates in two phases: At the outset, the judgment returned to BMI is that of the abstract qrels. The abstract qrels continue to be used until BMI identifies one document that is relevant not only according to the abstract qrels, but also according to the content qrels. Thereafter, the judgment returned to BMI is that of the content qrels.</p><p>In our pilot experiments, we found that the first method consistently yielded superior rank-based results, whether evaluated using the abstract qrels or the content qrels. The second method yielded consistently inferior results. The third method showed similar, but slightly inferior, results, to the first method, when evaluated using the content qrels. Based on our pilot results, we selected the first and third methods, denoted as Method A and Method B, respectively, for our official experiments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Stopping Criterion</head><p>For threshold-based evaluation, it was necessary to implement a stopping procedure to terminate screening when the best compromise between recall and effort had been achieved, for some definition of "best." In our opinion, technologyassisted review should be considered a satisfactory alternative to manual review only if it yields comparable or superior recall, with high probability. Toward this end, we deployed our knee method with default parameters (ρ = 156 − min(relret, 150), β = 100 <ref type="bibr" target="#b2">[3]</ref>), which interprets a sharp fall-off in the slope of the gain curve (recall vs. review effort) as evidence that substantially all relevant documents have been identified.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Runs and Evaluation</head><p>The Task 2 guidelines specify a plethora of run types and evaluation measures, which may be classified on two orthogonal dimensions:</p><p>1. Rank-based vs. threshold-based (or set-based) evaluation; and 2. Simple vs. cost-sensitive scoring.</p><p>The strategies to optimize these measures are incompatible, occasioning us to submit four versions of the output from each of our two runs, for a total of eight submissions, detailed in Table <ref type="table" target="#tab_0">1</ref>. The only difference between the "rank" and "thresh" runs is that the latter are truncated using the knee-method stopping procedure; the only difference between the "normal" and "cost" runs is that the "interaction field" "AF" is replaced by "AFS" where the document receives a "relevant" assessment, and by "AFN" where the document receives a "nonrelevant" assessment.</p><p>In 2015, we published the details and rationale for AutoTAR <ref type="bibr" target="#b0">[1]</ref>, which remains, to this date, the most effective TAR method of which we are aware. BMI implements AutoTAR exactly as described above, except for the substitution of Sofia-ML logistic regression in place of SVM light (see <ref type="bibr" target="#b3">[4,</ref><ref type="bibr">Section 3.1]</ref>). It has no dataset-or topic-specific tuning parameters; except for modifications to incorporate the CLEF corpora and relevance assessments, and our knee-method stopping procedure, we used BMI "out of the box."</p><p>The AutoTAR/BMI algorithm, as modified for CLEF, is detailed in Algorithm 1, which is reproduced from <ref type="bibr" target="#b0">[1]</ref> with the following changes:</p><p>-In Step 1, AutoTAR gives the option of starting with a relevant document, or with a synthetic document. Here, we used a synthetic document consisting of the title of the topic, and nothing else. -In Step 7, we introduced two different ways to simulate user feedback, corresponding to Method A and Method B, described above in Section 3.2. -In Step 10, we introduced the option to terminate the process when the knee-method stopping criterion was met.</p><p>Internally, BMI constructs a normalized TF-IDF ((1 + log tf ) • log N df ) wordvector representation of each document in the corpus (which, as noted in Section 3.1, consists of raw XML files), where a word is considered to be any sequence of two or more alphanumeric characters not containing a digit, that occurs at least twice in the corpus. Scoring is effected by Sofia-ML<ref type="foot" target="#foot_0">3</ref> with parameters "--learner type logreg-pegasos --loop type roc --lambda 0.0001 --iterations 200000." As noted above, these parameters were fixed when BMI was created in 2015.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>We present separately the results for our threshold-based and rank-based runs, reporting only simple threshold-based and simple rank-based measures for each, computed using the content qrels. At the time of writing, cost-sensitive evaluation was not available to CLEF participants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Threshold-Based Results</head><p>Our threshold-based results are shown in Table <ref type="table">2</ref>. Perhaps the most important result is shown in the first three lines: Across 30 topics, Method A identified all 607 articles referencing studies that should have been included, thus achieving 100% recall. Method B, on the other hand, identified 575 of the articles, achieving 97.9% recall. Method A, however, entailed the review of 79,765 (72.8%) of the 109,560 abstracts identified by the search phase, while method B entailed the review of only 52,934 (48.3%) of the documents.</p><p>In other words, Method A was more effective, but Method B was more efficient. According to the combined loss measure which considers both factors, Method B was superior.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Rank-Based Results</head><p>Our rank-based results are shown in Table <ref type="table">3</ref>. Work saved over sampling ("WSS")-a measure commonly reported for systematic review-reflects how many fewer documents would have been needed to have been reviewed to achieve a particular level of recall, if it were somehow known exactly when that level had been achieved. Thus, WSS, along with all other rank-based measures, is a measure of what might have been, rather than achieved effectiveness. According to WSS, Method A is marginally inferior to Method B at 95% recall (0.815 vs. 0.824), and at 100% recall (0.823 vs. 0.830).</p><p>Conversely, Method A is marginally superior to Method B in terms of the number of documents that had to be examined per topic before 100% recall was achieved (461 vs. 469, representing 12.6% and 12.8%, respectively, of the average number of documents per topic). In other words, Method A could have achieved 100% recall with roughly on-sixth the review effort, had a stopping procedure been able to determine when 100% recall had occurred. Similarly, Method B could have achieved 100% recall with roughly four times less effort that it actually required to achieve 97.9% recall, had a stopping procedure been available.</p><p>The Normalized Cumulative Gain ("NCG") results-which report the recall achieved when a specified fraction (between 10% and 100%) of the documents have been reviewed-tell much the same story: Very high recall could have been achieved at a fraction of the review effort, had it been know when high recall had been achieved.</p><p>In our opinion, cumulative measures like norm area and average precision yield very little insight into the actual or hypothetical effectiveness of technologyassisted review for screening purposes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Discussion</head><p>We believe that both sets of the CLEF assessments are incomplete with respect to the overall objective of identifying all studies that should be included in the review: The screening assessments are available only for documents retrieved by the search phase; the selection assessments are available only for documents retrieved by the search phase, and judged relevant during the screening phase. Therefore, from the assessments, it is impossible to determine whether an article not retrieved by the search phase, or an article eliminated during the screening phase, describes a study that should have been included in the review. The Task 2 architecture tacitly assumes that no such articles exist; in other words, that the search and screening phases used to generate the relevance assessments were infallible, and each attained 100% recall.</p><p>Such an assumption is unrealistic, and limits the recall of any simulated TAR method to that of the manual review to which it is compared <ref type="bibr" target="#b5">[6]</ref>. As noted in the Cochrane Handbook <ref type="bibr" target="#b8">[9]</ref> with regard to the search phase: "[T]here comes a point where the rewards of further searching may not be worth the effort required to identify the additional references." And with regard to the screening phase: "Using at least two authors may reduce the possibility that relevant reports will be discarded <ref type="bibr">(Edwards 2002 [7]</ref>)."</p><p>Our hypothesis that our TAR runs found relevant articles that were missed by the search phase, or incorrectly discarded in the screening phase, is based on results from other domains <ref type="bibr" target="#b5">[6]</ref>, where TAR acting as a "second assessor" was able to identify potentially relevant documents that had been judged "nonrelevant" by a human assessor. When we applied Method A to the 30 topics, it identified 9,250 potentially relevant articles for which the abstract qrel was "not relevant." Acquiring a second opinion on each of these documents would increase the cost of the TAR review by approximately 12%, and would, we believe, yield a substantial number of relevant documents, over and above the 670 identified in the abstract qrels.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Official Waterloo CLEF Task 2 Submissions.</figDesc><table><row><cell cols="4">Run Name Method Rank/Threshold Simple/Cost Sensitive</cell></row><row><cell>A-rank-cost</cell><cell>A</cell><cell>Rank</cell><cell>Cost Sensitive</cell></row><row><cell>A-rank-normal</cell><cell>A</cell><cell>Rank</cell><cell>Simple</cell></row><row><cell>A-thresh-cost</cell><cell>A</cell><cell>Threshold</cell><cell>Cost Sensitive</cell></row><row><cell>A-thresh-normal</cell><cell>A</cell><cell>Threshold</cell><cell>Simple</cell></row><row><cell>B-rank-cost</cell><cell>B</cell><cell>Rank</cell><cell>Cost Sensitive</cell></row><row><cell>B-rank-normal</cell><cell>B</cell><cell>Rank</cell><cell>Simple</cell></row><row><cell>B-thresh-cost</cell><cell>B</cell><cell>Threshold</cell><cell>Cost Sensitive</cell></row><row><cell>B-thresh-normal</cell><cell>B</cell><cell>Threshold</cell><cell>Simple</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>See https://www.nlm.nih.gov/bsd/pmresources.html.</figDesc><table><row><cell>1 Available</cell><cell>under</cell><cell>GNU</cell><cell>General</cell><cell>Public</cell><cell>License</cell><cell>at</cell></row><row><cell cols="3">http://cormack.uwaterloo.ca/trecvm.</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table><note>2 </note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">See https://github.com/glycerine/sofia-ml.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Autonomy and reliability of continuous active learning for technology-assisted review</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1504.06868</idno>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Waterloo (Cormack) participation in the TREC 2015 Total Recall Track</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Twenty-Fourth Text REtrieval Conference, TREC 2015</title>
				<meeting>The Twenty-Fourth Text REtrieval Conference, TREC 2015<address><addrLine>Gaithersburg, Maryland, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">November 17-20, 2015, 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Engineering quality and reliability in technology-assisted review</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2016</title>
				<meeting>the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2016<address><addrLine>Pisa, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">July 17-21, 2016. 2016</date>
			<biblScope unit="page" from="75" to="84" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Scalability of continuous active learning for reliable high-recall text classification</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, CIKM 2016</title>
				<meeting>the 25th ACM International on Conference on Information and Knowledge Management, CIKM 2016<address><addrLine>Indianapolis, IN, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">October 24-28, 2016. 2016</date>
			<biblScope unit="page" from="1039" to="1048" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">When to stop&quot; Waterloo (Cormack) participation in the TREC 2016 Total Recall Track</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Twenty-Fifth Text REtrieval Conference, TREC 2016</title>
				<meeting>The Twenty-Fifth Text REtrieval Conference, TREC 2016<address><addrLine>Gaithersburg, Maryland, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">November 15-18, 2016, 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Navigating imprecision in relevance assessments on the road to total recall: Roger and me</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2017</title>
				<meeting>the 40th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2017<address><addrLine>Tokyo, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">August 7-11, 2017, 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records</title>
		<author>
			<persName><forename type="first">P</forename><surname>Edwards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Clarke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Diguiseppi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pratap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wentz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistics in Medicine</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="1635" to="1640" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">TREC 2016 Total Recall Track overview</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roegiest</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Twenty-Fifth Text REtrieval Conference, TREC 2016</title>
				<meeting>The Twenty-Fifth Text REtrieval Conference, TREC 2016<address><addrLine>Gaithersburg, Maryland, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">November 15-18, 2016, 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Higgins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Green</surname></persName>
		</author>
		<title level="m">Cochrane handbook for systematic reviews of interventions</title>
				<imprint>
			<publisher>John Wiley &amp; Sons</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="volume">4</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of the CLEF technologically assisted reviews in empirical medicine</title>
		<author>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Azzopardi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Spijker</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2017 -Conference and Labs of the Evaluation forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">September 11-14, 2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">TREC 2015 total recall track overview</title>
		<author>
			<persName><forename type="first">A</forename><surname>Roegiest</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Grossman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Clarke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Twenty-Fifth Text REtrieval Conference, TREC 2015</title>
				<meeting>The Twenty-Fifth Text REtrieval Conference, TREC 2015<address><addrLine>Gaithersburg, Maryland, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">November 17-20, 2015, 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Overview of the CLEF ehealth evaluation lab</title>
		<author>
			<persName><forename type="first">H</forename><surname>Suominen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zuccon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R M</forename><surname>Palotti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -8th International Conference of the CLEF Association, CLEF 2017</title>
		<title level="s">Proceedings, Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017-09-11">2017. September 11-14, 2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
