<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Ranking Abstracts to Identify Relevant Evidence for Systematic Reviews: The University of Sheffield&apos;s Approach to CLEF eHealth 2017 Task 2 Working Notes for CLEF 2017</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Amal</forename><surname>Alharbi</surname></persName>
							<email>ahalharbi1@sheffield.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Sheffield</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mark</forename><surname>Stevenson</surname></persName>
							<email>mark.stevenson@sheffield.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Sheffield</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Ranking Abstracts to Identify Relevant Evidence for Systematic Reviews: The University of Sheffield&apos;s Approach to CLEF eHealth 2017 Task 2 Working Notes for CLEF 2017</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">89DBCD365210F5ADB8D71F2D420538F6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:30+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes Sheffield University's submission to CLEF 2017 eHealth Task 2: Technologically Assisted Reviews in Empirical Medicine. This task focusses on the identification of relevant evidence for systematic reviews in the medical domain. Participants are provided with systematic review topics (including title, Boolean query and set of PubMed abstracts returned) and asked to identify the abstracts that provide evidence relevant to the review topic. Sheffield University participated in the simple evaluation. Our approach was to rank the set of PubMed abstracts returned by the query by making use of information in the topic including title and Boolean query. Ranking was based on a simple TF.IDF weighted cosine similarity measure. This paper reports results obtained from six runs: four submitted to the official evaluation, an additional run and a baseline approach.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Systematic reviews attempt to identify, synthesise and summarise evidence available to answer a research question. They form the backbone of evidence-based approaches to medicine where they are used to answer complex questions such as "How effective are statins for heart attack survivors?" <ref type="bibr" target="#b0">[1]</ref>.</p><p>The process of creating a systematic review is time-consuming with a single review often requiring 6 to 12 months of effort from expert reviewers <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>. Text mining techniques have been shown to be a useful way to reduce this effort <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>. CLEF eHealth Task 2 "Technologically Assisted Reviews in Empirical Medicine" focusses on the application of text mining to the process of developing systematic reviews with the aim to reduce the effort required.</p><p>This paper is organised as follows: Section 2 introduces CLEF eHealth Task 2. Section 3 describes our approach to this task. Section 4 discusses the results obtained from applying this approach to both the development and test datasets. Finally, Section 5 presents the conclusions and potential future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task Description</head><p>The process of identify relevant evidence for a systematic review usually consists of multiple stages <ref type="bibr" target="#b7">[8]</ref>:</p><p>1. Boolean Search: Experts construct a boolean query designed to identify all evidence relevant to the review question. This query is run against a medical database such as PubMed and set of titles and abstracts returned. 2. Title and Abstract Screening: Experts screen the titles and abstracts retrieved to identify those that are potentially relevant for inclusion in the review. 3. Document Screening: The full document content is then retrieved for any title and abstract that has been identified as being relevant in the previous stage. These are then examined in a second round of expert screening to form a final decision about their relevance to the review.</p><p>In CLEF eHealth 2017 <ref type="bibr" target="#b8">[9]</ref>, Task 2 <ref type="bibr" target="#b7">[8]</ref> focuses on the second stage of systematic review (Title and Abstract Screening). Participants are required to develop methods to rank a list of PubMed abstracts returned by a boolean query (stage 1) so that relevant documents appear as early as possible.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Method</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Datasets</head><p>Participants are provided with two datasets: a development set and a test set. The development dataset contains 20 topics and the test dataset contains 30 topics. All reviews focus on Diagnostic Test Accuracy (DTA). The queries were manually constructed by expert reviewers from the Cochrane collaboration<ref type="foot" target="#foot_0">1</ref> . For each topic, participants are provided with topic id, review title, boolean query and a list of PubMed documents identifiers retrieved by the query. The collection contains a total of 266,967 abstracts.</p><p>Figure <ref type="figure">1</ref> shows examples of two topics from the development dataset. Two different formulations were used for the Boolean queries: OVID and PubMed. The queries are generally complex and contain multiple operators. Table <ref type="table" target="#tab_0">1</ref> shows operators commonly used in both types of query <ref type="bibr" target="#b2">[3]</ref>.</p><p>Participants also provided with files that indicate which of the titles and abstracts returned by the Boolean query were indicated as being relevant after the Title and Abstract Screening and Document Screening stages (see Section 2), referred to as the abstract qrels and content qrels respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">University of Sheffield's Approach</head><p>The University of Sheffield's submission to Task 2 ranked the list of PubMed abstracts retrieved for each topic with the intention of returning relevant ones as early as possible. The approach is completely automatic since queries are processed algorithmically and without manual intervention<ref type="foot" target="#foot_1">2</ref> . In addition, relevance feedback is not used.</p><p>Our method makes use of three pieces of information from the topic: (1) the title, (2) terms extracted from the Boolean query and (3) MeSH terms extracted from the Boolean query. Information for (2) and (3) are extracted from the Boolean query using a simple parser designed to interpret both OVID and PubMed style queries. Terms  Figure <ref type="figure">2</ref> shows examples of terms extracted from the query for topic CD008643 (see Figure <ref type="figure">1</ref>). Some MeSH terms (e.g. Spine) are also standard English words that could appear as a term in an abstract. To avoid false matches all MeSH terms extracted from a query are prefixed with the string Mesh. In addition, MeSH terms are preprocesssed to remove whitespace and punctuation (e.g. Lumbar vertibrae becomes MeshLumbarvertibrate). Example MeSH terms extracted from the same query are shown in Figure <ref type="figure">3</ref>. The abstracts returned by the Boolean query for each topic defined as the list of PMIDs (PubMed identifier) provided with the topic are downloaded from PubMed <ref type="foot" target="#foot_2">3</ref> . The text of the title, abstract and MeSH terms are extracted and the MeSH terms preprocessed using the same approach that was applied to the Boolean query.</p><p>Pre-processing is applied to both the PubMed abstracts and information extracted from the topics. The text is tokenised, converted to lower case, stop words/punctuation are removed and the remaining tokens stemmed <ref type="foot" target="#foot_3">4</ref> .</p><p>The information extracted from the topic and each of the abstracts are converted into tf.idf-weighted vectors. The similarity between the topic and each of the abstracts is then generated by computing the cosine metric for the pair of vectors <ref type="foot" target="#foot_4">5</ref> . Abstracts are ranked based on this similarity score.</p><p>Results are output in the TREC format shown in Table <ref type="table" target="#tab_1">2</ref> where:</p><p>-TOPIC-ID: topic identifier provided by CLEF 2017.</p><p>-INTERACTION: this field is assigned the value NF in all our runs to indicate that relevance feedback is not used -PID: PubMed document identifier -RANK: rank of the document according to the cosine similarity score -SCORE: cosine similarity score described above -RUN-ID: run identifier</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Runs</head><p>Four runs were officially submitted for the official evaluation: Sheffield-run-1, Sheffieldrun-2, Sheffield-run-3, and Sheffield-run-4. In addition, a baseline run (Sheffield-baseline) and additional approach (Sheffield-run-5) were also implemented and evaluated. A description of each run is presented below. -Sheffield-baseline In this run the list of PubMed abstracts are randomly ordered. This is intended to represent the scenario in which the results of the Boolean query are simply evaluated in the order in which they are retrieved without any attempt to identify those most likely to be relevant. This situation simulates common practise within many systematic review projects in which reviewers examine each of the retrieved abstracts in turn. The score of each abstract is calculated using the following equation:</p><formula xml:id="formula_0">score = n − r + 1 n (<label>1</label></formula><formula xml:id="formula_1">)</formula><p>where n is the total number of abstracts returned by the Boolean query and r the abstract's rank in the random ordering. -Sheffield-run-1 Abstracts returned by the Boolean query are ranked by comparing them against only the topic title. -Sheffield-run-2 Abstracts are compared with the topic title and terms extracted from the Boolean query. -Sheffield-run-3 Abstracts are compared with the topic title and both terms and MeSH terms extracted from the Boolean query. -Sheffield-run-4 This run is the same as Sheffield-run-2 except that the PubMed stop-words list <ref type="bibr" target="#b11">[12]</ref> is used rather than the one from sklearn. -Sheffield-run-5 Abstracts are compared against the topic title and MeSH terms extracted from the Boolean query. (This run is the same as Sheffield-run-3 except that terms extracted from the Boolean query are not included when computing the similarity.)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results and Discussion</head><p>Task 2 consists of two formal evaluations: simple evaluation and cost-effective evaluation. The University of Sheffield participated only in the simple evaluation setup and did not attempt to optimise the approach for the cost-effective evaluation. Evaluation was carried out using the script provided by the task organisers<ref type="foot" target="#foot_5">6</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Development Dataset</head><p>The development dataset contains of 20 DTA topics (see Section 3.1). Tables <ref type="table" target="#tab_3">3 and 4</ref> present the results for the approaches described in Section 3.3 applied to this dataset for the abstract and content qrels respectively.</p><p>As expected, all of the implemented methods outperform the simple baseline approach. This demonstrates that even straightforward ranking techniques provide potential benefit to systematic reviewers by ensuring that documents more likely to be relevant are placed higher in the rankings. We have previously demonstrated a similar results for a single systematic review <ref type="bibr" target="#b4">[5]</ref> and that finding is supported by these results which represent a substantially larger dataset.</p><p>The best result of the submitted runs for the abstract qrels (Table <ref type="table" target="#tab_2">3</ref>) was achieved by Sheffield-run-4 which achieved the average precision (ap) score of 0.223, an improvement of 0.173 against the baseline. It also achieved the best results for work saved over sampling (wss) and area under the cumulative recall curve normalized by the optimal area (norm_area) metrics. It is also close to the best result for the average of the minimum number of abstracts returned to retrieve all relevant ones (last_rel) metric.</p><p>For the content qrels (Table <ref type="table" target="#tab_3">4</ref>), both Sheffield-run-4 and Sheffield-run-5 are strong. Sheffield-run-4 produced the best scores for last_rel and norm_area and close to the best result of wss. Sheffield-run-5 achieved the best score for ap and wss_95.</p><p>Results from the development dataset suggest that including terms extracted from the Boolean query is beneficial (e.g. compare Sheffield-run-1 and Sheffield-run-2). However, the usefulness of MeSH terms extracted is less clear. Performance decreases when these are added to the title and query terms (e.g. compare Sheffield-run-2 and Sheffield-run-3). Results are mixed when they are used instead of query terms (e.g. compare Sheffield-run-1 and Sheffield-run-5), there is no improvement for the abstract evaluation but some benefit for the content evaluation. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Test Dataset</head><p>The development dataset contains of 30 DTA topics (see Section 3.1). Tables <ref type="table" target="#tab_5">5 and 6</ref> show the results for the abstract and content qrels respectively. The highest ap scores were achieved using Sheffield-run-2 and Sheffield-run-4 for both the abstract and content qrels (Tables <ref type="table" target="#tab_5">5 and 6</ref>). The overall pattern of results suggest that Sheffield-run-4 is the best performing run on the test data.</p><p>Results from the development and test datasets indicate the strong relative performance of Sheffield-run-4. This indicates that including terms extracted from Boolean query and using the PubMed stop-words list are benefical for this task. There were some relevant documents in the test data set for which our approach assigned a score of 0 and this caused NCG@100 scores to be less than 1. This was observed at both the content and abstract level for the development and test datasets. The scoring script treats these documents as not being included in the ranking. The problem could be resolved by adding a small delta value to each score.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion and Future Work</head><p>This paper described the University of Sheffield's approach to CLEF 2017 Task 2. Information from the review title and Boolean query was used to rank the abstracts returned by the query using standard similarity measures. The title and terms extracted from the Boolean query were found to be the most useful information for this task. All of the submitted runs outperform a baseline approach based on random ordering.</p><p>In future we plan to refine the techniques for extracting terms and MeSH terms from the Boolean query (Section 3.2) by taking account of the query structure and MeSH hierarchy. We also plan to develop techniques to minimise the cost of identifying relevant evidence and make use of ActiveLearning to improve the ranking based on feedback from reviewers.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>4 Figure 1 .</head><label>41</label><figDesc>Figure 1. Example topics from Cochrane reviews used in development dataset [10,11].</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>'Figure 2 .Figure 3 .</head><label>23</label><figDesc>Figure 2. Sample of terms extracted from the query of topic CD008643</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>OVID and PubMed common query operators</figDesc><table><row><cell>OVID</cell><cell></cell></row><row><cell>/ or .sh.</cell><cell>MeSH terms</cell></row><row><cell>.mp.</cell><cell>MeSH subheading</cell></row><row><cell>.tw.</cell><cell>Text words</cell></row><row><cell>.ti,ab.</cell><cell>Title/abstract</cell></row><row><cell>PubMed</cell><cell></cell></row><row><cell cols="2">[mesh] or [mh] MeSH terms</cell></row><row><cell>[sh]</cell><cell>MeSH subheading</cell></row><row><cell>[tw]</cell><cell>Text words</cell></row><row><cell>[tiab]</cell><cell>Title/abstract</cell></row></table><note>and MeSH terms modified by certain operators (e.g. not and adj) are not extracted.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Sample output for Sheffield-run-1</figDesc><table><row><cell cols="2">TOPIC-ID INTERACTION</cell><cell>PID</cell><cell cols="2">RANK SCORE</cell><cell>RUN-ID</cell></row><row><cell>CD010438</cell><cell>NF</cell><cell cols="2">18388501 17</cell><cell>0.245 Sheffield-run-1</cell></row><row><cell>CD010438</cell><cell>NF</cell><cell cols="2">16884987 18</cell><cell>0.239 Sheffield-run-1</cell></row><row><cell>CD010438</cell><cell>NF</cell><cell cols="2">22164456 19</cell><cell>0.238 Sheffield-run-1</cell></row><row><cell>CD010438</cell><cell>NF</cell><cell cols="2">22193152 20</cell><cell>0.236 Sheffield-run-1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Results of runs evaluated against development dataset using abstract qrels</figDesc><table><row><cell>RUN-ID</cell><cell>ap</cell><cell cols="3">last_rel wss_100 wss_95 norm_area</cell></row><row><cell cols="3">Sheffield-baseline 0.05 7121.65</cell><cell>0.036</cell><cell>0.033</cell><cell>0.495</cell></row><row><cell cols="3">Sheffield-run-1 0.188 5793.7</cell><cell>0.138</cell><cell>0.385</cell><cell>0.815</cell></row><row><cell cols="3">Sheffield-run-2 0.223 4449.65</cell><cell>0.184</cell><cell>0.434</cell><cell>0.836</cell></row><row><cell cols="3">Sheffield-run-3 0.217 4768.85</cell><cell>0.17</cell><cell>0.415</cell><cell>0.83</cell></row><row><cell cols="3">Sheffield-run-4 0.223 4496.85</cell><cell>0.188</cell><cell>0.442</cell><cell>0.839</cell></row><row><cell cols="3">Sheffield-run-5 0.182 5866.6</cell><cell>0.135</cell><cell>0.344</cell><cell>0.808</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Results of runs evaluated against development dataset using content qrels</figDesc><table><row><cell>RUN-ID</cell><cell>ap</cell><cell cols="4">last_rel wss_100 wss_95 norm_area</cell></row><row><cell cols="2">Sheffield-baseline 0.01</cell><cell>6575.3</cell><cell>0.104</cell><cell>0.077</cell><cell>0.465</cell></row><row><cell cols="3">Sheffield-run-1 0.094 2204.95</cell><cell>0.574</cell><cell>0.61</cell><cell>0.855</cell></row><row><cell cols="3">Sheffield-run-2 0.104 2097.2</cell><cell>0.549</cell><cell>0.589</cell><cell>0.867</cell></row><row><cell cols="3">Sheffield-run-3 0.095 2141.35</cell><cell>0.533</cell><cell>0.593</cell><cell>0.859</cell></row><row><cell cols="3">Sheffield-run-4 0.107 1999.35</cell><cell>0.568</cell><cell>0.611</cell><cell>0.875</cell></row><row><cell cols="3">Sheffield-run-5 0.108 2701.7</cell><cell>0.545</cell><cell>0.615</cell><cell>0.855</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>Results of runs evaluated against test dataset using abstract qrels</figDesc><table><row><cell>RUN-ID</cell><cell>ap</cell><cell cols="3">last_rel wss_100 wss_95 norm_area</cell></row><row><cell cols="4">Sheffield-baseline 0.045 3727.433 0.039</cell><cell>0.031</cell><cell>0.483</cell></row><row><cell cols="3">Sheffield-run-1 0.17 2678.333</cell><cell>0.31</cell><cell>0.422</cell><cell>0.818</cell></row><row><cell cols="3">Sheffield-run-2 0.218 2441.7</cell><cell>0.385</cell><cell>0.493</cell><cell>0.845</cell></row><row><cell cols="4">Sheffield-run-3 0.199 2404.967 0.384</cell><cell>0.473</cell><cell>0.841</cell></row><row><cell cols="4">Sheffield-run-4 0.218 2382.467 0.395</cell><cell>0.488</cell><cell>0.847</cell></row><row><cell cols="3">Sheffield-run-5 0.158 2650.8</cell><cell>0.303</cell><cell>0.423</cell><cell>0.809</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 .</head><label>6</label><figDesc>Results of runs evaluated against test dataset using content qrels</figDesc><table><row><cell>RUN-ID</cell><cell>ap</cell><cell cols="2">last_rel wss_100 wss_95 norm_area</cell></row><row><cell cols="3">Sheffield-baseline 0.023 3307.793 0.088</cell><cell>0.067</cell><cell>0.478</cell></row><row><cell cols="3">Sheffield-run-1 0.12 1801.724 0.517</cell><cell>0.544</cell><cell>0.844</cell></row><row><cell cols="3">Sheffield-run-2 0.176 1928.828 0.534</cell><cell>0.58</cell><cell>0.87</cell></row><row><cell cols="3">Sheffield-run-3 0.153 1902.586 0.524</cell><cell>0.588</cell><cell>0.866</cell></row><row><cell cols="3">Sheffield-run-4 0.177 1846.586 0.543</cell><cell>0.587</cell><cell>0.874</cell></row><row><cell cols="3">Sheffield-run-5 0.114 1922.103 0.487</cell><cell>0.541</cell><cell>0.836</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://www.cochrane.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">The approach was implemented using Python v3.6</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">The Entrez package from biopython.org was used.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">NLTK's tokenize and LancasterStemmer packages are used for tokenisation and stemming. The list of stop words provided by scikit-learn (scikit-learn.org/stable/) is used for most runs.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4"><ref type="bibr" target="#b4">5</ref> Scikit-learn's TfidfVectorizer and linear_kernel packages were used for these steps</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://github.com/leifos/tar</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">An Introduction to Systematic Reviews</title>
		<author>
			<persName><forename type="first">D</forename><surname>Gough</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Oliver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thomas</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Sage</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A Prospective Evaluation of an Automated Classification System to Support Evidence-based Medicine and Systematic Review</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ambert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mcdonagh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AMIA Annual Symposium Proceedings</title>
				<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="volume">2010</biblScope>
			<biblScope unit="page" from="121" to="125" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Boolean Versus Ranked Querying for Biomedical Systematic Reviews</title>
		<author>
			<persName><forename type="first">S</forename><surname>Karimi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pohl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Scholer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cavedon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zobel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">BMC medical informatics and decision making</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Reducing Systematic Review Workload Through Certainty-based Screening</title>
		<author>
			<persName><forename type="first">M</forename><surname>Miwa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>O'mara-Eves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ananiadou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Biomedical Informatics</title>
		<imprint>
			<biblScope unit="volume">51</biblScope>
			<biblScope unit="page" from="242" to="253" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Identifying Potential Early Biomarkers of Acute Myocaridal Infarction in the Biomedical Literature: A Comparison of Text Mining and Manual Sifting Techniques</title>
		<author>
			<persName><forename type="first">S</forename><surname>Paisley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sevra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stevenson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Archer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Preston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chilcott</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Society for Pharmacoeconomics and Outcomes Research (ISPOR) 19th Annual European Congress</title>
				<meeting>the International Society for Pharmacoeconomics and Outcomes Research (ISPOR) 19th Annual European Congress<address><addrLine>Vienna, Austria</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Use of Cost-effectiveness Analysis to Compare the Efficiency of Study Identification Methods in Systematic Reviews</title>
		<author>
			<persName><forename type="first">I</forename><surname>Shemilt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thomas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Systematic reviews</title>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Using Text Mining for Study Identification in Systematic Reviews</title>
		<author>
			<persName><forename type="first">A</forename><surname>O'mara-Eves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mcnaught</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Miwa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ananiadou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Systematic reviews</title>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">CLEF Technologically Assisted Reviews in Empirical Medicine Overview</title>
		<author>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Azzopardi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Spijker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2017 -Conference and Labs of the Evaluation forum, CEUR Workshop Proceedings</title>
				<meeting><address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2017">September 11-14 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">CLEF 2017 eHealth Evaluation Lab Overview</title>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Suominen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Robert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Spijker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Palotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zuccon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF 2017 -8th Conference and Labs of the Evaluation Forum</title>
		<title level="s">Lecture Notes in Computer Science (LNCS</title>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017-09">September,, 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Imaging Modalities for the Non-invasive Diagnosis of Endometriosis</title>
		<author>
			<persName><forename type="first">V</forename><surname>Nisenblat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">M</forename><surname>Bossuyt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Farquhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Hull</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cochrane Database of Systematic Reviews</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">D009591</biblScope>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Red Flags to Screen for Vertebral Fracture in Patients Presenting with Low-back Pain</title>
		<author>
			<persName><forename type="first">C</forename><surname>Williams</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Henschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Maher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Van Tulder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Macaskill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Irwig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cochrane Database of Systematic Reviews</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">D008643</biblScope>
			<date type="published" when="2013">2013. 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">table, stopwords</title>
		<ptr target="https://www.ncbi.nlm.nih.gov/books/NBK3827/table/pubmedhelp.T.stopwords/" />
	</analytic>
	<monogr>
		<title level="m">-pubmed help -ncbi bookshelf</title>
				<imprint>
			<date type="published" when="2017-05-07">7 May 2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
