=Paper= {{Paper |id=Vol-1180/CLEF2014wn-eHealth-Verberne2014 |storemode=property |title=A Language-modelling Approach to User-Centred Health Information Retrieval |pdfUrl=https://ceur-ws.org/Vol-1180/CLEF2014wn-eHealth-Verberne2014.pdf |volume=Vol-1180 |dblpUrl=https://dblp.org/rec/conf/clef/Verberne14 }} ==A Language-modelling Approach to User-Centred Health Information Retrieval== https://ceur-ws.org/Vol-1180/CLEF2014wn-eHealth-Verberne2014.pdf
    A language-modelling approach to User-Centred
             Health Information Retrieval

                                     Suzan Verberne

                     Institute for Computing and Information Sciences
                                Radboud University Nijmegen
                                     s.verberne@cs.ru.nl



        Abstract. In this working notes paper we present our methodology
        and the results we obtained in task 3a of the CLEF eHealth lab 2014. In
        the set-up of our experiments we assumed that the discharge summary
        provides the context of the patient’s query, and therefore may contain
        useful background information that can be used to retrieve more rele-
        vant results. The central component in our approach is the Indri search
        engine, with its Language Modelling-based retrieval model. We experi-
        mented with query expansion using terms extracted from the discharge
        summary, and using terms extracted from the UMLS thesaurus. We
        obtained a small positive effect from expanding the Indri queries with
        terms from both sources. Our future work is directed at improving our
        term extraction and query expansion strategies.


1     Introduction
In this working notes paper we present our methodology and the results we obtained
in task 3a of the CLEF eHealth lab 20141 . The aim of the eHealth track is “to
evaluate systems that support laypeople in searching for and understanding their
health information” [1, 3]. The goal of task 3 is “to provide valuable and relevant
documents to patients, so as to satisfy their health-related information needs” [2].
The situation that the task simulates is a patient who has learned from his physician
what his diagnosis is, and then starts searching the internet for medical information
about his illness. The physician’s information about the patient has been registered
in the patient’s discharge summary.
    The data that was distributed for task 3 consists of:
 – the 2012 crawl of approximately one million medical documents made available
   by the EU-FP7 Khresmoi project2 in plain text form;
 – a set of English ‘layman’ queries that individuals may realistically pose based
   on the content of their discharge summaries. The query creators used the disor-
   der diagnosed in the discharge summary as main query term. The training set
   contained 5 queries and the test set 50;
 – (optionally, after following human subjects training and obtaining the certifi-
   cate) the collection of 299 discharge summaries from task 2 among which the
   discharge summaries belonging to the train and test queries for task 3.
   In the set-up of our experiments we assumed that the discharge summary pro-
vides the context of the patient’s query, and therefore may contain useful back-
ground information that can be used to retrieve more relevant results. Our research
question is:
1
    http://clefehealth2014.dcu.ie/task-3
2
    http://www.khresmoi.eu/
                                              269
      Does query expansion with terms from the medical context of a patient’s
      query lead to better results?
   The central component in our approach is the Indri search engine3 , with its
Language Modelling-based retrieval model. We experiment with query expansion
using terms extracted from the discharge summary, and using terms extracted from
the UMLS thesaurus.

2     Our approach
2.1    Data preprocessing
Document collection We preprocessed the document collection by splitting the
corpus files in separate documents and saving for each document its uid, date, url
and content.

Discharge summaries We obtained the corpus of 299 discharge summaries that
was distributed for CLEF-eHealth task 2. We processed all discharge summaries that
are referred to by a query in the query set from task 3. We cleaned the discharge
summaries from all variables of the form [** ...**] (e.g. [**MD Number 2860**]),
from abbreviations (using the regular expression [a-z]\.([a-z]\.)+), and from
numbered lists (sequences of lines matching the regular expression ^[0-9]+\..*$).

2.2    Indexing and Retrieval
We used the Indri API to index the CLEF collection and set up a query interface to
the index. We applied a stopword list to the CLEF collection at indexing time. As
ranking model, we use the Indri LM with Dirichlet smoothing and Pseudo-Relevance
feedback (PRF) using the Ponte Expander. As parameters for the PRF, we used:
number of feedback docs: 20, number of feedback terms: 3. We did not optimize
these parameters for the current task but used the optimal settings from a previous
task.

2.3    Query construction
All characters that are not alphanumeric, no hyphen or whitespace are removed
from the query and all letters are lowercased. The words in the query are concate-
nated into one string and combined using the combine function in the Indri query
language. We used two types of queries: short queries, consisting of a concatenation
of the title and description fields, and long queries, consisting of a concatenation of
all four fields title, description, profile and narr. For example, for this query:

       QTRAIN2014.1
       08114-027513-DISCHARGE SUMMARY.txt
       
       MRSA and wound infection
       What is MRSA infection and is it dangerous?
       This 60 year old lady has had coronary artery bypass
       grafting surgery and during recovery her wound has been infected.
       She wants to know how dangerous her infection is, where she got
       it and if she can be infected again with it.
       Documents should contain information about sternal wound
       infection by MRSA. They should describe the causes and the
       complications.
3
    http://www.lemurproject.org/indri/
                                         270
Table 1. The top-5 informative terms that we extracted from the discharge summaries
belonging to training queries 2, 3 and 4 using the method by [4]

QTRAIN2014.2         QTRAIN2014.3            QTRAIN2014.4
mcg                  unit stay               aortic dissection
mute                 medical intensive care lake
thrush               end-stage renal disease type
levothyroxine sodium department by           nodule
mesylate             saturating              right kidney



the following short query is created: #combine(mrsa and wound infection what
is mrsa infection and is it dangerous)


2.4     Query expansion with terms from discharge summaries

We used the discharge summaries to create more informative queries, covering some
of the medical background information of the patient. To this end, we aimed to
extract the most relevant terms from the discharge summary belonging to each
query. We extracted all n-grams with n = {1, 2, 3} from the discharge summary
and ranked them by their relevance. For calculating the relevance of each n-gram,
we used Kullback-Leibler divergence for informativeness and phraseness [4]. In this
method, term relevance is based on the expected loss between two language models,
measured with point-wise Kullback-Leibler divergence. Tomokiyo and Hurst propose
to mix two models for term scoring:

 – phraseness (how tight are the words in the multi-word term):

                                                       P (t)
                              kldivP = P (t) ∗ log Qn                                (1)
                                                     i=1 (P (ui ))

   in which P (ui ) is the probability of the ith unigram inside the n-gram t
 – informativeness (how informative is the term for the foreground corpus):

                                                          P (t)f g
                                kldivI = P (t)f g ∗ log                              (2)
                                                          P (t)bg

      in which P (t)f g is the probability of the n-gram t in a foreground collection and
      P (t)bg is the probability of t in the background collection.

For expanding query q, we used the discharge summary belonging to q as foreground
collection, and all 299 discharge summaries in the task2 corpus as background col-
lection. The parameter γ is the weight of the informativeness score relative to the
phraseness score:

                    T ermRelevance = γ ∗ kldivI + (1 − γ) ∗ kldivP                   (3)
We used γ = 0.9, giving a higher weight to informativeness than to phraseness. We
sort the n-grams by their T ermRelevance. Table 1 shows the top-5 terms extracted
for a few training queries. Note that not all terms are qualitative, some seem too
generic to give information about this specific patient. This is caused by the relative
sparseness of the data from which the terms are extracted, only one document.
    In the runs that use the discharge summaries (run2–4), we added the top-k
(k = {2, 5}) of terms extracted from the discharge summary to the query, again
using unstructured queries with the #combine-operator. This implies that multi-
word terms are treated as separate words in the query, not as phrases.
                                               271
    C UMG6-MRSA-METHICILLIN-RESISTANT-STAPHYLOCOCCUS-AUREUS-INFECTION: *0 #{{{
       S *0
          MRSA - Methicillin resistant Staphylococcus aureus infection *0
          Methicillin-resistant staphylococcus aureus (MRSA) *0
       .
    -#}}}

    C UMG9-METHICILLIN-RESISTANT-STAPHYLOCOCCUS-AUREUS: *0 #{{{
       S *0
          Methicillin resistant Staphylococcus aureus *0
          methicillin resistant Staphylococcus aureus *0
          MRSA *0
       .
    -#}}}


Fig. 1. Two thesaurus entries from UMLS for the query title “MRSA and wound infection”


2.5    Query expansion with terms from a thesaurus

We use thesaurus expansion to enricht the queries with non-personalized informa-
tion. The idea is that the patient might use layman’s terminology and that the
technical-medical synonyms of these terms might give better retrieval results. For
this purpose, we use an old version of the ULMS-thesaurus that we have stored
locally. In the UMLS, medical terms are ordered in synonym sets. An example is
shown in Figure 1. We looked up all query titles from the train and test set in the
ULMS: First we preprocessed the query titles so that only the first content word is
kept (e.g. “MRSA and wound infection” becomes “MRSA”). Then we processed all
synonym sets that

 – contain one of the preprocessed query titles, or
 – in which a word from the query title that is between 3 and 5 uppercase char-
   acters (presumably an abbreviation) is used between brackets (e.g. Methicillin-
   resistant staphylococcus aureus (MRSA))

For each query title, we created one expansion set by merging all synonym sets in
which the preprocessed query title occurs, disregarding synonym sets with 30 or
more terms in them (because generic synonym sets such as “UMG2-BODY-PART-
ORGAN-OR-ORGAN-COMPONENT” contain dozens and sometimes hundreds of
terms). From the expansion set, we removed duplicates and near-duplicates (the
only difference is a plural -s, hyphens and underscores, or capitalization). Then
we sort the expansion terms by the number of times they occur in synonym sets
together with the query title and select the top-k terms to be added to the query.
   In the runs that use the thesaurus (run2,3,5,6), we added the top-k (k = 5) of
terms extracted from the thesaurus to the query, again using unstructured queries
with the #combine-operator. This implies that multi-word terms are treated as
separate words in the query, not as phrases.


3     Submitted runs

 1. Baseline: Indri LM retrieval with Pseudo-Relevance feedback (PRF). Queries are
    unstructured combination of title and description, using the #combine- operator
    in the Indri query language
 2. Same as Run1, and:
                                         272
Table 2. Evaluation of our runs in terms of Precision (P), nDCG, MAP and the number
of relevant results retrieved (ret rel). For each evaluation measure (column), the highest
scoring run is marked in boldface.

Run ID                P@10           nDCG@10        MAP           ret rel
NIJM EN Run.1         0.5740         0.5708         0.3036        2330
NIJM EN Run.2         0.6180         0.6149         0.2825        2190
NIJM EN Run.3         0.5960         0.5772         0.2606        2154
NIJM EN Run.4         0.5680         0.5669         0.2695        2176
NIJM EN Run.5         0.5880         0.5773         0.2609        2165
NIJM EN Run.6         0.5220         0.5302         0.2180        1939
NIJM EN Run.7         0.5220         0.5302         0.2180        1939


      – We expanded each query with terms from the discharge summary (k=5)
      – We expanded each query with maximally 5 terms from the UMLS thesaurus
 3. Same as Run2, but with k=2 for the terms from the discharge summary
 4. Same as Run2 (k=5), but without the thesaurus expansion. Note that in our
    original submission, there was an error in this run, as a result of which its results
    were equal to Run 3. We reproduced the run and included the correct results
    in the results section below.
 5. Same as Run1, and:
      – We expanded each query with maximally 5 terms from the UMLS thesaurus
 6. Same as Run5 but with long queries: unstructured combination of title, descrip-
    tion, profile and narrative
 7. Same as Run6 but without the thesaurus expansion

4     Results
4.1   Evaluation of the runs
Table 2 shows a summary of the results obtained for our runs. From the results, we
make the following observations:
 – Run 1, with short queries and no query expansion, gives the best results in
   terms of MAP and the number of retrieved and relevant results.
 – Run 2, with 5 terms from the discharge summary and maximum 5 terms from
   the thesaurus, gives the best results in terms of nDCG@10 and P@10. This
   suggests that a combination of terms from the discharge summary and from the
   thesaurus added to the query can lead to better results in the top-10.
 – Run 3, with 2 expansion terms from the discharge summary and maxumum 5
   terms from the thesaurus, gives worse results than Run 2, with 5 terms from
   the discharge summary. This suggests that the most informative terms from the
   discharge summaries are not always ranked at top; or that a combination of
   terms is needed to retrieve more relevant results.
 – Run 4, with 5 terms from the discharge summary and no terms from the the-
   saurus, does not beat the baseline Run 1. This suggests that without the the-
   saurus terms, the discharge summary terms do not lead to improvement.
 – Run 5, with maximally 5 terms from the thesaurus and no terms from the
   discharge summary, gives almost the same results as Run 3, with 2 terms from
   the discharge summary. This suggests that if there is relevant information in
   the discharge summary terms, it only leads to improvement if we add enough
   terms (2 is not enough).
 – Run 6, with long queries, no terms from the discharge summary and maximally
   5 terms from the thesaurus, gives worse results than Run 5, with short queries
   and maximally 5 terms from the thesaurus. This suggests that short queries are
   better than long queries.
                                              273
Fig. 2. Per-query results for Precision@10 for our best run (of the runs submitted), Run2.
For each query, the height of a bar represents the gain/loss of our Run (grey) and the best
Run (white) over the median run of all submitted runs.


 – Run 7, with long queries and no thesaurus expansion gives the exact same results
   as Run 6, with long queries and maximally 5 terms from the thesaurus. This
   suggests that when using the long queries with the fields ‘profile’ and ‘narrative’
   fully included, makes it senseless to add a few thesaurus terms. Since runs 6
   and 7 are the poorest of all our runs, using the longer queries suspectedly leads
   to more irrelevant results than the shorter queries. Even if the thesaurus terms
   are informative, they cannot compensate this.

4.2   Per-query analysis
Figure 2 shows the per-query results for our best run (of the runs submitted), Run2.
It shows a large divergence between queries. For some queries (the ones where there
is no white bar), our run (the grey bar) scores the best of all runs, but for others,
our run scores far below median. In follow-up work we will investigate the query
characteristics that cause our method to be more or less successful. We suspect that
the success depends at least partly on the quality of the extracted additional terms.


5     Conclusions
Since we did not do comprehensive analyses of our results, we can only draw prelim-
inary conclusions. We worked with a Language Modelling approach, and although
the Indri retrieval model can handle long queries (it will find the best possible
matches for the combination all query terms — not necessarily all terms are present
in the retrieved documents), we found that expanding the query consisting of the
fields title and description with the fields profile and narrative has a negative effect
on the retrieval results. However, we did get a positive effect from adding the 5
most informative terms from the discharge summary and maximally 5 synonyms
from the thesaurus to the queries.
                                           274
    In the near future, we plan to do more analyses in order to find out what factors
play a role in the success of query expansion. Our current line of work focuses
on personalization of IR through terms extracted from personal documents. The
discharge summary is a good example of a document that may provide additional
information about the context of a user’s query. For that purpose, we aim to improve
our term extraction and query expansion strategies. In addition, we will investigate
how thesaurus expansion could be applied successfully (selecting the query terms
to look up, selecting the synonym sets to expand with).


Acknowledgements
This publication was supported by the Dutch national program COMMIT (project
P7 SWELL).


References
1. Lorraine Goeuriot, G Jones, Liadh Kelly, Johannes Leveling, Allan Hanbury, Henning
   Müller, Sanna Salanterä, Hanna Suominen, and Guido Zuccon. Share/clef ehealth
   evaluation lab 2013, task 3: Information retrieval to address patients questions when
   reading clinical reports. Online Working Notes of CLEF, CLEF, 2013.
2. Lorraine Goeuriot, Liadh Kelly, Wei Li, Joao Palotti, Pavel Pecina, Guido Zuccon,
   Allan Hanbury, Gareth Jones, and Henning Mueller. Share/clef ehealth evaluation lab
   2014, task 3: User-centred health information retrieval. In Proceedings of CLEF 2014,
   2014.
3. Liadh Kelly, Lorraine Goeuriot, Hanna Suominen, Tobias Schrek, Gondy Leroy,
   Danielle L. Mowery, Sumithra Velupillai, Wendy W. Chapman, David Martinez, Guido
   Zuccon, and Joao Palotti. Overview of the share/clef ehealth evaluation lab 2014. In
   Proceedings of CLEF 2014, Lecture Notes in Computer Science (LNCS). Springer, 2014.
4. Takashi Tomokiyo and Matthew Hurst. A language model approach to keyphrase ex-
   traction. In Proceedings of the ACL 2003 workshop on Multiword expressions: analysis,
   acquisition and treatment-Volume 18, pages 33–40. Association for Computational Lin-
   guistics, 2003.




                                             275