=Paper=
{{Paper
|id=Vol-1179/CLEF2013wn-CLEFeHealth-ZucconEt2013b
|storemode=property
|title=Retrieval of Health Advice on the Web AEHRC at ShARe/CLEF eHealth Evaluation Lab Task 3
|pdfUrl=https://ceur-ws.org/Vol-1179/CLEF2013wn-CLEFeHealth-ZucconEt2013b.pdf
|volume=Vol-1179
|dblpUrl=https://dblp.org/rec/conf/clef/ZucconKN13
}}
==Retrieval of Health Advice on the Web AEHRC at ShARe/CLEF eHealth Evaluation Lab Task 3==
Retrieval of Health Advice on the Web AEHRC at ShARe/CLEF eHealth Evaluation Lab Task 3 G. Zuccon1 , B. Koopman1,2 , A. Nguyen1 1 The Australian e-Health Research Centre (CSIRO), Brisbane, Australia 2 Queensland University of Technology, Brisbane, Australia {guido.zuccon, bevan.koopman, anthony.nguyen}@csiro.au Abstract. This paper details the participation of the Australian e- Health Research Centre (AEHRC) in the ShARe/CLEF 2013 eHealth Evaluation Lab – Task 3. This task aims to evaluate the use of infor- mation retrieval (IR) systems to aid consumers (e.g. patients and their relatives) in seeking health advice on the Web. Our submissions to the ShARe/CLEF challenge are based on language models generated from the web corpus provided by the organisers. Our baseline system is a standard Dirichlet smoothed language model. We enhance the baseline by identifying and correcting spelling mistakes in queries, as well as expanding acronyms using AEHRC’s Medtex medical text analysis platform. We then consider the readability and the author- itativeness of web pages to further enhance the quality of the document ranking. Measures of readability are integrated in the language models used for retrieval via prior probabilities. Prior probabilities are also used to encode authoritativeness information derived from a list of top-100 consumer health websites. Empirical results show that correcting spelling mistakes and expanding acronyms found in queries significantly improves the effectiveness of the language model baseline. Readability priors seem to increase retrieval effectiveness for graded relevance at early ranks (nDCG@5, but not pre- cision), but no improvements are found at later ranks and when consid- ering binary relevance. The authoritativeness prior does not appear to provide retrieval gains over the baseline: this is likely to be because of the small overlap between websites in the corpus and those in the top-100 consumer-health websites we acquired. 1 Introduction Patients usually have limited medical knowledge and thus patient education can improve a patients understanding of their health condition, as well as adherence and compliance to a treatment. The use of web search engines to retrieve medi- cal advice online is increasingly popular [1–3]. The ShARe/CLEF 2013 eHealth Evaluation Lab Task 3 [4] aims to evaluate search engines in the task of retriev- ing health advice on the web, and to uncover issues that the research community can address to improve the effectiveness of search engines technologies for this task. The Australian e-Health Research Centre (AEHRC) contributed 4 runs to this year’s challenge. Our methods are based on the language modelling frame- work for information retrieval [5]. Our baseline submission (teamAEHRC.1.3) implements a language model with Dirichlet smoothing [6]. The remaining sub- missions build upon this baseline approach. Specifically, we consider the contri- bution to retrieval effectiveness of query error correction and acronym expansion (teamAEHRC.5.3), readability measures (teamAEHRC.6.3), authoritativeness as derived from a list of top-100 consumer health websites (teamAEHRC.7.3). Our approaches are detailed in Section 2. Empirical results obtained over the web corpus compiled by the ShARe/CLEF 2013 eHealth Evaluation Lab Task 3 organisers highlight the importance of cor- recting typographic errors in health consumer queries, as well as normalising acronyms to their expanded form to increase the quality of the query represen- tation. Considering the readability of web pages when providing information to health consumers provides improvements in retrieval effectiveness when consid- ering the graded relevance results at early ranks. Web page authority as assessed from a list of top-100 consumer-health websites does not seem to improve re- trieval quality: this may be due to the limited overlap between the website list and the ShARe/CLEF 2013 document corpus. Details of the results achieved by our submissions are given in Section 3. 2 Methods The next sections describes the methods we used to address the problem of re- trieving web pages for health consumers seeking for medical advice [3]. The doc- ument rankings generated by our methods were submitted to the ShARe/CLEF 2013 eHealth Evaluation Lab Task 3. All methods are implemented using the Lemur/Indri information retrieval toolkit1 . 2.1 A baseline Language Model (teamAEHRC.1.3) We used a language modelling approach with Dirichlet smoothing as the base- line retrieval method. Following this approach, documents are ranked according to the probability of a document d given the submitted query Q, i.e. P (d|Q), computed as: Y P (qi |d) + µP (qi |C) P (d|Q) ≈ P (Q|d)P (d) ≈ (1) |d| + µ qi ∈Q where the prior probability P (d) is considered uniform over the document col- lection and can thus be ignored for ranking equivalence reasons, |d| is the length in tokens of document d, P (qi |C) is the maximum likelihood estimate of qi in the collection, and µ is the Dirichlet smoothing parameter. This parameter was set to 2,500 in all our submission; this is a common value for the smoothing 1 http://lemurproject.org/ parameter. In our implementation (as it is common in IR), only documents that contain at least one of the query terms are considered for retrieval for each given query. Of these, only the top 1,000 documents, according to Equation 1, are used to form the submission. 2.2 Correcting spelling mistakes and expanding acronyms (teamAEHRC.5.3) The analysis of the training set provided by the task organisers revealed that queries may contain (i) spelling mistakes or alternative spellings, e.g., groupo B for group B; (ii) acronyms and abbreviations, e.g., Cdiff for Clostridium difficile. Similar cases can in fact also be found in the test set, e.g. Hypothyreoidism for Hypothyroidism, ASA for acetylsalicylic acid (aspirin)2 . The presence of spelling mistakes and the use of acronyms in queries may adversely affect retrieval of a standard keyword-based system like ours based on language modelling. To overcome this issue, we use the AEHRC’s Medtex medical text analysis platform [7] to individuate misspelled terms (and uncommon variants of medical terms), as well as acronyms. Medtex is part of the medical natural language processing technology that the AEHRC uses to deliver automated solutions to improve health service delivery, like cancer incidence and mortality reporting [8, 9] and radiology reconciliation [10]. To correct candidate misspellings and uncommon variants individuated by Medtex, we implemented a call to the Google web search engine3 and extracted the query correction suggestion (i.e. “Showing results for”) provided by the search engine. To expand candidate acronyms individuated by Medtex, we parsed the list of common abbreviations used in medical prescriptions provided by Wikipedia4 . This list contains triples, where abbreviation is the target acronym or abbreviation expression, latin is the Latin term that represents the abbreviation (if any) and meaning is the English expansion. If an English expansion was available for an abbreviation, then we ignored the latin term, otherwise we used the latin term as a translation of the abbreviation. When spelling corrections and acronym expansions are produced for candi- date terms of a query, we create a new query formulation that appends to the original query terms those from the spelling correction and acronym expansion. The new query formulation is then used to score documents against using the model of Equation 1. This forms the submission named teamAEHRC.5.3. 2 Note, however, that the test set was not consulted when developing the approach described here. 3 http://www.google.com 4 http://en.wikipedia.org/wiki/List_of_abbreviations_used_in_medical_ prescriptions 2.3 Taking readability into account: the readability prior (teamAEHRC.6.3) Health consumers seeking medical advice on the web do not usually have exper- tise in the medical domain and are thus not familiar with the medical language. A page providing health information for health practitioners (e.g., doctors and nurses) is likely to be difficult to read for a health consumer, such as a patient. We follow this intuition and argue that web pages retrieved for providing advice to health consumer need to be easily understood by a non-expert reader. We thus enhance the approach used for building the submission teamAEHRC.5.3 (pre- vious section) by considering document readability. We use a common measure of text readability to estimate how likely the content of a web page is under- standable by health consumers. The selected readability measure is the Flesch Kincaid Reading Ease (FRES) formula [11]. This measure provides a score be- tween 0 and 100. A high score indicates that the text is easy to read, while low scores suggest the text is complicated to understand. The Flesch Kincaid Reading Ease measure has been used in previous work on readability of health content, for example, to assess whether informed consent forms for participation in oncology research are readable by patients and their families [12]. The Flesch Kincaid Reading Ease is calculated according to the following formula: #(words) #(syllable) 206.835 − 1.015 · − 84.6 · (2) #(sentences) #(words) where the function #(x) provides the total count of item x in the document, e.g. #(syllable) is the total number of syllable in the document. To consider the readability measure during the retrieval process, we com- pute a prior probability distribution over all documents in the collection, where the value of the prior probability assigned to a document is proportional to its Flesch Kincaid Reading Ease score. Thus, documents that are more readable according to the Flesch Kincaid Reading Ease measure would be more likely relevant according to our prior. The prior is integrated in the retrieval formula by modifying Equation 1 so that the readability prior is substituted to the uni- form prior, P (d), used for the previous runs. This method forms the submission named teamAEHRC.6.3. 2.4 Considering authoritativeness: a prior for the top-100 consumer health websites (teamAEHRC.7.3) Health information presented to consumers should not only be easy to under- stand, but also reliable. To take reliability of the content into account dur- ing the retrieval process, we obtained a list of recommended health-consumer web sites. The list has been compiled by CAPHIS5 and can be retrieved at http://caphis.mlanet.org/consumer/. This list contains 100 sites that have 5 The Consumer and Patient Health Information Section (CAPHIS) is part of the Medical Library Association, an association of health information professionals. been selected according to criteria that include currency, credibility, content and audience. We see these criteria as an overall measure of how authoritative the websites are. To integrate authoritativeness information in the retrieval process, we took an approach similar to that used for the document prior in run teamAEHRC.6.3. An uniform prior was computed for all documents in the collection. For documents whose base URL is in the CAPHIS list, we boosted the corresponding prior by 10 times. The resulting score distribution was then normalised to resemble a probability distribution, this formed our authoritativeness prior. The prior was then applied together with the readability prior, transforming the retrieval formula to the following: Y P (qi |d) + µP (qi |C) P (d|Q) ≈ P (Q|d)Pr (d)Pa (d) ≈ Pr (d)Pa (d) (3) |d| + µ qi ∈Q where Pr (d) is the readability prior for document d and Pa (d) is the authori- tativeness prior. This formed the method used for generating teamAEHRC.7.3 submission. 3 Evaluation on the ShARe/CLEF 2013 Challenge 3.1 Evaluation Settings Details of the collection used are given by Goeuriot et al. [3]. We indexed the document collection using Lemur; the INQUIRY stop list and the Porter stemmer as implemented in Lemur were used when indexing documents. Table 1. Percentage indicating the coverage of the relevance assessments with re- spect to the top 10 results retrieved by each of our submissions for each query in the ShARe/CLEF 2013 retrieval challenge. A low percentage for a submission suggests that metric such as P@10 and nDCG@10 may be underestimated for that submission. Run id % coverage teamAEHRC.1.3 100.00% teamAEHRC.5.3 100.00% teamAEHRC.6.3 71.40% teamAEHRC.7.3 43.60% Runs are evaluated according to the guidelines provided by the ShARe/CLEF 2013 eHealth Evaluation Lab Task 3 organisers; Precision@10 and nDCG@10 are used as main evaluation measures. Organisers formed the pools used for relevance assessments by considering, for each query, the top 10 documents re- trieved by only selected participants submissions. The selected submission in- clude our teamAEHRC.1.3 and teamAEHRC.5.3 runs; thus the set of top 10 documents retrieved for each query by teamAEHRC.6.3 and teamAEHRC.7.3 may contain unjudged documents. Unjudged documents are considered irrele- vant (according to standard IR practice); this may result in an underestimation of the Precision@10 and nDCG@10 for teamAEHRC.6.3 and teamAEHRC.7.3. Table 1 reports the percentage of the top 10 rank positions of each query for our submissions that are covered by the ShARe/CLEF 2013 relevance assessments. The table highlights that the top 10 rankings of each queries for submissions teamAEHRC.6.3 and teamAEHRC.7.3 have only partially been assessed: less than half of the documents retrieved by teamAEHRC.7.3 in the top 10 ranks have been judged. Thus, metrics such as precision and nDCG calculated at rank 10 for these submissions may be underestimating the quality of the submission itself. The extent to which the effectiveness of these runs are underestimated de- pends upon the breadth of the pools assessed in the ShARe/CLEF 2013 eHealth Evaluation Lab Task 3. Table 2. Official retrieval effectiveness of our approaches obtained on the ShARe/CLEF 2013 eHealth Evaluation Lab Task 3. (a) Retrieval effectiveness measured (b) Retrieval effectiveness measured by nor- by Precision at rank 5 and 10 (P@5, malised discounted cumulative gain at rank P@10). 5 and 10 (nDCG@5, nDCG@10). Run id P@5 P@10 Run id nDCG@5 nDCG@10 teamAEHRC.1.3 0.4440 0.4540 teamAEHRC.1.3 0.3825 0.3991 teamAEHRC.5.3 0.4560 0.4840 teamAEHRC.5.3 0.3946 0.4223 teamAEHRC.6.3 0.4440 0.4220 teamAEHRC.6.3 0.4128 0.3992 teamAEHRC.7.3 0.2080 0.2200 teamAEHRC.7.3 0.1937 0.1992 3.2 Results Results of our submissions are summarised in Tables 2(a) and 2(b) for Preci- sion@10 and nDCG@10 respectively. The results suggest that expanding queries by including spelling corrections, common alternative spelling of medical terms, acronym and abbreviation expan- sions (submission teamAEHRC.5.3), does sensibly improve retrieval effectiveness across the whole set of evaluation metrics. The inclusion of readability measures as a prior of the language model used for retrieval does not provide the hoped improvement when considering binary relevance and precision. When graded relevance is considered, accounting for the readability of the content does pro- vide better effectiveness at early ranks, as measured by the increase of nDCG@5 obtained by teamAEHRC.6.3 over the baseline teamAEHRC.1.3 (+8%) and tea- mAEHRC.5.3 (+5%). These gains are, however, not maintained at lower ranks. The inclusion of the authoritativeness prior actually degrades the retrieval ef- fectiveness; this is likely to be a result of the limited overlap between the URLs in the CAPHIS list and those in the ShARe/CLEF 2013 corpus. An additional factor influencing the low performance of this run may be the limited overlap be- tween retrieved and assessed documents. As discussed in Section 3.1, this factor may cause an underestimation of the retrieval effectiveness of this submission, the extent of which can only be judged based on the completeness and breadth of the pool of documents and systems that have been assessed. (a) teamAEHRC.1.3 (b) teamAEHRC.5.3 (c) teamAEHRC.6.3 (d) teamAEHRC.7.3 Fig. 1. Gains and losses in precision at 10 of our submissions with respect to the median system at ShARe/CLEF 2013. White bars identify the gains of the best system for each query. Figure 1 reports the gains and the losses in precision at 10 of our submis- sions when compared to the median system at ShARe/CLEF 2013. These are compared to the highest gains measured on a query-basis in the ShARe/CLEF 2013 challenge. Submission teamAEHRC.5.3 provides large gains over teamAEHRC.1.3 (and the median of ShARe/CLEF 2013 systems) in queries where error corrections and acronym expansion are fundamental. For example, the system used for tea- mAEHRC.1.3 retrieved only two documents for query qtest1, ‘hypothyreoidism’ ; this is because only two documents contain this uncommon spelling of this con- dition. Our query terms correction method added to the query the term ’hy- pothyroidism’, which allowed for the retrieval of a large quantity of relevant doc- uments. The submission formed by the system implementing our query terms correction (and acronym expansion) technique, i.e. teamAEHRC.5.3, performed best for queries like query qtest1. Overall, the gains in retrieval effectiveness over the baseline provided by the technique used for teamAEHRC.5.3 are partially lost when introducing the readability prior in the retrieval method. However, the query-by-query analysis of Figure 1 highlights that losses do not affect all queries: the introduction of the readability prior delivers effectiveness gains for some of the queries in the ShARe/CLEF 2013 test set. This is the case for ex- ample for queries qtest10, ‘dysplasia and multiple sclerosis’, and qtest11, ‘chest pain and liver transplantation’. In the latter case, the system that uses the read- ability prior for retrieval results is, in fact, the best across all ShARe/CLEF 2013 systems. Finally, we have observed that our submissions teamAEHRC.6.3 have no re- sults for qtest50, and teamAEHRC.7.3 has no results for qtest49 and qtest50. An analysis of our code revealed an error in the generation of the final docu- ment rankings to submit to the challenge; this produced truncated results for submissions teamAEHRC.6.3 and teamAEHRC.7.3. 4 Conclusions This paper outlined our submissions to the ShARe/CLEF 2013 eHealth Evalu- ation Lab Task 3. Our approaches are based on a Dirichlet smoothing language modelling framework, and investigate the effect of misspelling corrections and acronyms expansions in queries, and the inclusion of readability and authorita- tiveness information in the scoring function. A preliminary analysis of our results have highlighted the gains that can be obtained by correcting misspellings in queries and expanding acronyms and abbreviations. The inclusion of readability information have shown promise for retrieving information for health consumers seeking medical advice on the web. Further analysis is required to better gauge the impact of our approaches. Future work will also investigate alternative ap- proaches for (i) computing readability of health content, (ii) encoding readability in the retrieval function. References 1. Fox, S.: Health topics: 80% of internet users look for health information online. Technical report, Pew Research Center (February 2011) 2. White, R., Horvitz, E.: Cyberchondria: Studies of the escalation of medical con- cerns in web search. Technical report, Microsoft Research (2008) 3. Goeuriot, L., Kelly, L., Jones, G., Zuccon, G., Suominen, H., Hanbury, A., Muller, H., Leveling, J.: Creation of a new medical information retrieval evaluation bench- mark targeting patient information needs. In: 5th International Workshop on Evaluating Information Access (EVIA). (2013) 4. Suominen, H., Salanterä, S., Velupillai, S., Chapman, W.W., Savova, G., Elhadad, N., Mowery, D., Leveling, J., Goeuriot, L., Kelly, L., Martinez, D., Zuccon, G.: Overview of the ShARe/CLEF eHealth Evaluation Lab 2013. In: CLEF 2013. Lecture Notes in Computer Science (LNCS), Springer (2013) 5. Ponte, J.M., Croft, W.B.: A language modeling approach to information retrieval. In: Proceedings of the 21st annual international ACM SIGIR conference on Re- search and development in information retrieval, ACM (1998) 275–281 6. Zhai, C., Lafferty, J.: A study of smoothing methods for language models applied to information retrieval. ACM Transactions on Information Systems (TOIS) 22(2) (2004) 179–214 7. Nguyen, A., Moore, J., Lawley, M., Hansen, D., Colquist, S.: Automatic extraction of cancer characteristics from free-text pathology reports for cancer notifications. In: Health Informatics Conference. (2011) 117–124 8. Zuccon, G., Nguyen, A., Bergheim, A., Wickman, S., Grayson, N.: The impact of ocr accuracy on automated cancer classification of pathology reports. In: Health Informatics: Building a Healthcare Future Through Trusted Information-Selected Papers from the 20th Australian National Health Informatics Conference (Hic 2012). Volume 178., Ios PressInc (2012) 250 9. Butt, L., Zuccon, G., Nguyen, A., Bergheim, A., Grayson, N.: Classification of cancer-related death certificates using machine learning. Australasian Medical Journal 6(5) (2013) 292–299 10. Zuccon, G., Wagholikar, A., Nguyen, A., Chu, K., Martin, S., Lai, K., Greenslade, J.: Automatic classification of free-text radiology reports to identify limb fractures using machine learning and the snomed ct ontology. In: AMIA Clinical Research Informatics. (2013) 11. Kincaid, J.P., Fishburne Jr, R.P., Rogers, R.L., Chissom, B.S.: Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Technical report, DTIC Document (1975) 12. Grossman, S.A., Piantadosi, S., Covahey, C.: Are informed consent forms that describe clinical oncology research protocols readable by most patients and their families? Journal of Clinical Oncology 12(10) (1994) 2211–2215