<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>KISTI at CLEF eHealth 2015 Task 2</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Heung-Seon Oh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yuchul Jung</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kwang-Young Kim</string-name>
          <email>glorykim@kisti.re.kr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Korea Institute of Science and Technology Information</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Laypeople (e.g., patients and their caregivers) usually use queries which describe a sign, symptom or condition to obtain relevant medical information on the Web. They can fail to find useful information for diagnosing or understanding their health conditions because the search results delivered by existing medical search engines do not fit the information needs of users. To deliver useful medical information, we attempted to combine multiple ranking methods, explicit semantic analysis (ESA), a cluster-based external expansion model (CBEEM), and concept-based document centrality (CBDC), using external medical resources to improve retrieval performance. As a first step, initial documents are searched using a baseline method. Based on the initial documents, ranking methods are selectively applied. Our experiments with combinations of ranking methods aim to find the best means of computing accurate similarity scores using different external medical resources. The best performance was obtained when the CBEEM and the CBDC were used together.</p>
      </abstract>
      <kwd-group>
        <kwd>medical information retrieval</kwd>
        <kwd>external expansion model</kwd>
        <kwd>concept-based retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The general public searches the Web to acquire medical information to diagnose their
symptoms and find related health information. Unfortunately, searchers such as
laypeople without medical knowledge can fail to find the necessary information in a search
query because they are often not only unfamiliar with medical terminology but also
uncertain about their exact questions. Tackling queries for laypeople has been a
challenging issue with regard to medical information retrieval (IR) because existing Web
search engines often fail to deliver satisfactory search results because the required
information is not properly understood. To mitigate the difficulties of laypeople (e.g.,
patients and their relatives), Conference and Labs of the Evaluation Forum (CLEF)
launched the eHealth Evaluation Lab [4]. Specifically, Task 2 of
        <xref ref-type="bibr" rid="ref6">CLEF 2015</xref>
        eHealth
[10] explores circumlocutory queries consisting of the signs and symptoms of a medical
condition.
      </p>
      <p>As a participant in task 2, this paper introduces a re-ranking framework which
attempts to combine selectively different ranking components, such as explicit semantic
analysis (ESA), a cluster-based external expansion model (CBEEM), and
conceptbased document centrality (CBDC). The main goal of our framework is an accurate
estimation of the similarity score by combining different ranking methods using
external medical resources.</p>
      <p>Within our re-ranking framework, a query-likelihood method with Dirichlet
smoothing as a baseline was utilized to obtain the initial document set.   is re-ranked with
the help of ranking components using external medical resources, two biomedical
collections (i.e., TREC CDS [11] and OHSUMED [5]) and ICD-10 1extracted from
Wikipedia. In our experiments, we designed eight runs which combine more than one
reranking components, except run 1, which represents the baseline. Among the eight runs,
the best performance was observed in runs 6 and 8, when the CBEEM and the CBDC
were combined. The best performances, in runs 6 and 8, were 0.3864 (P@10) and
0.3464 (NDCG@10).</p>
      <p>The rest of this paper is organized as follows. Section 2 presents our ranking
framework in detail. The experimental results are described in Section 3. Section 4 concludes
with a short summary.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <sec id="sec-2-1">
        <title>Re-ranking framework</title>
        <p>The key idea of our method is to devise a re-ranking framework which estimates an
accurate similarity score between a query and a document using external medical
resources. To do this, we build a pool of re-ranking components with external resources.
Figure 1 shows an overview of our re-ranking framework. For a given query Q, a set of
documents,    = { 1,  2, … ,   }, is retrieved from collection C using a search
engine. In this paper, a query-likelihood method with Dirichlet smoothing (QLD) [14] is
utilized to obtain   . Then, we focus on re-ranking   using external resources to
improve the performance. Specifically, two biomedical collections, TREC CDS and
OHSUMED, and ICD-10 as extracted from Wikipedia were used as external resources.
Based on   , re-ranking is performed through a series of ranking components in the
pool.
1 http://apps.who.int/classifications/icd10/browse/2015/en
   
( ,  ) = exp (−</p>
        <p>(  ||  ))
= exp ( − ∑  ( |  ) 

 ( |  )
 ( |  )
)
where   and   are the query and document unigram language models,
respecKLD has been attractive because effective pseudo-relevance feedback methods have
been proposed to estimate more accurate query language models in an effort to improve
performance. The research questions are how to estimate accurate query and document
language models to improve the retrieval performance.</p>
        <p>In general, a query model is estimated by maximum likelihood estimation MLE), as
shown below:
2.2</p>
        <sec id="sec-2-1-1">
          <title>Basic Foundation</title>
          <p>Before explaining the details of the three different re-ranking components, we introduce
the basic foundation of the language modeling framework for IR to provide a deeper
explanation. In language modeling for IR, the KL-divergence method (KLD) is a
popular scoring function to compute similarity scores by estimating unigram language
models for a query Q and a document D [6, 7, 9]:
(1)
(2)
(3)
tively.
where  ( ,  ) is the count of a word w in query Q and | | is the number of words in
A document model is estimated using Dirichlet smoothing to avoid zero probabilities
and to improve the retrieval performance through an accurate estimation [14]:
where  ( ,  ) is the count of a word w in document D,  ( | ) is the probability of
a word w in collection C, and  is the Dirichlet prior parameter.</p>
          <p>Query expansion aims to reveal information needs not expressed in Q by adding
more useful words. Pseudo-relevance feedback (PRF) is a popular query expansion
approach to update a query. Updating a query with PRF assumes that the top-ranked
documents  = { 1,  2, … ,  | |} in the initial search results relevant to a given query and
the words in F are useful to modify a query for a better representation. A relevance
model (RM) serves to estimate a multinomial distribution  ( | ), which is the
likelihood of a word w in query Q. The first version of the relevance model (RM1) is defined
as follows:
  1( | ) = ∑  ( |  ) (  | )
 ∈
 ∈
 ∈
= ∑  ( |  )
 ( |  ) (  )</p>
          <p>( )
∝ ∑  ( |  ) (  ) ( |  )
(4)</p>
          <p>RM1 is composed of three components: the document prior  (  ), the document
weight  ( |  ), and the term weight in a document  ( |  ). In general,  (  ) is
assumed to have a uniform distribution without knowledge of document D.  ( |  ) =
∏ ∈  ( |  ) ( , ) indicates the query-likelihood score.  ( |  ) can be estimated
using various smoothing methods, such as Dirichlet-smoothing. Various strategies are
applicable to estimate these components.</p>
          <p>To improve the retrieval performance, a new query model can be estimated by
combing the relevance model and the original query model. RM3 [1] is a variant of a
relevance model which is used here to estimate a new query model with RM1,
 ( | ′ ) = (1 −  ) ⋅  ( |  ) +  ⋅   1( | ),
(5)
where  is a control parameter between the original query model and the feedback
model.
2.3</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>Re-ranking Components</title>
          <p>Component 1 - Explicit Semantic Analysis: Concept-based IR using an explicit
semantic analysis (ESA) [3] is a well-known approach used to deal with a vocabulary
mismatch problem between a query and a document, where the words in the query and
document are mapped to concepts. In medical IR, methods [2, 12] employ MetaMap to
map words to concepts in the Unified Medical Language System (UMLS). Processing
millions of documents in a collection using MetaMap involves a considerable amount
of time complexity. To avoid this difficulty, concepts relevant to International
Classification Diseases (ICD-10) were used as a concept resource because they are closely
related to diseases. These concepts were collected from Wikipedia. Articles linked to
the name of the section and the sub-section of ICD-10 were crawled. As a result, 3,784
articles with 93,756 unique words were obtained. The title of an article was used as a
medical concept. Figure 2 shows an example of the medical concept Bubonic plague 2
in Wikipedia. Based on the concepts, a word-concept matrix filled with standard
TFIDF values was constructed. Then, a similarity score between a query and a document
is computed after concept mapping, as shown in Figure 3. Cosine similarity was utilized
as a scoring function.</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2 http://en.wikipedia.org/wiki/Bubonic_plague</title>
        <p>Component 2 - Cluster-based External Expansion Model: There are several
medical collections, TREC CDS and OHSUMED, available to researchers, as medical
collections have been developed for different purposes. For re-ranking purposes, these
collections can be used as textual resources to build more robust external expansion
models [13]. To this end, we revised an existing external expansion model (EEM) by
combining it with a cluster-based document model [8]. The key idea of the EEM is to
generate a feedback model by determining the proper contributions of multiple
collections for a given query. Formally, the EEM is defined as follows:
 
( | ) ∝ ∑  ( |  ) ⋅  (  ) ∑  ( |  ) ⋅  ( |  ) ⋅  ( |  ).
(6)
 ∈</p>
        <p>∈</p>
        <p>Specifically, the EEM consists of five components: the prior collection probability,
document relevance, collection relevance, document importance, and word probability.
Prior collection probability  (  ) is the prior importance of a collection among all the
collections in use. Without the prior knowledge of collections, it can be ignored by
setting a uniform probability  (  ) = |1|. Document relevance  ( |  ) is the
relevance of a document D to a given query Q. Precisely, it is a query-likelihood score
given to a document. Collection relevance  ( |  ) is the relevance of a query Q with
respect to a collection C. This component determines the query-dependent contribution
of a collection when constructing the EEM. To avoid time-consuming iteration over a
collection C, it can be estimated using the most highly relevant documents with the
assumption that documents are equally important in a given collection C. Thus, it is the
| |
| | +</p>
        <p>| | + 
= (1 −   ) ⋅ [
 ( | ) +
 ( | )] +   ⋅  ( | ),
where λE is a control parameter for all collections in E.</p>
        <p>Our CBEEM is defined by revising  ( |  ) in Equation 6 and replacing it with that
of Equation 7. Based on this revision, the CBEEM is expected to be a probability
distribution over topical words because it is combined with individual RMs owing to the
decrease in the probability of common words in the feedback documents. Then, a new
query model is estimated with the CBEEM as follows:</p>
        <p>| |
average score of the feedback documents in  
. Document importance  ( |  )
refers to the importance of a document D in a collection C. This is also ignored by setting
to a uniform probability  ( |  ) = 1 without the prior knowledge of documents in a
collection C. Word probability  ( |  ) is a probability of observing a word w in a
document D. In [13], the MLE is utilized to estimate this component.</p>
        <p>In the cluster-based document model [8], a document model is smoothed with cluster
and collection models in which the clusters are generated with the K-means algorithm.
Therefore, we can obtain more accurate document models because the probabilities of
words which occur frequently in a cluster or a collection are decreased. Similarly, we
can assume that each collection corresponds to a cluster explicitly partitioned over E.
This assumption allows the use of the cluster-based document model without any
additional computations with K-means clustering, as K is determined via | |, and each
collection is a cluster. All that is required is to utilize the statistics of a collection C for a
cluster. Then, a document model is defined as follows:
 ( |  ) = (1 −   ) ⋅
 ( | ′ ) = (1 −  ) ⋅  ( |  ) +  ⋅  
( | )</p>
        <p>Component 3 - Concept-based Document Centrality: To utilize external resources,
we designed a concept-based document centrality method (CBDC) as an additional
reranking component. The key idea originated from centrality-based document scoring,
which utilizes the associations among documents in the search results [6]. The
centralities are computed through two steps - similarity matrix construction and a
randomwalk step. Among the initial documents, implicit links are generated because there are
no explicit links among them. Then, the documents are re-ranked by combining the
initial and centrality scores, as follows:

e( ,  ) = 


( ,  ) ⋅ 

 ( ,  )</p>
        <p>However, the CBDC differs from previous approaches [6] in two aspects. First, we
attempted to capture the associations among a query and documents explicitly when
computing document centralities, while the previous method only considered the
associations among documents. Second, the CBDC captures the associations at the concept
level while the previous method focused on the word level. The CBDC is estimated as
follows. First, the document-concept weight matrix is constructed by concept mapping.
In this matrix, the query is augmented at the ends of the rows. Then, a
document-document similarity matrix is computed using the document-concept weight matrix. Due
to the need to augment the query, the CBDC considers the associations of documents
with respect to a query. Next, a random walk was performed to compute centrality
scores. We only utilized the centrality scores of documents.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <sec id="sec-3-1">
        <title>Data</title>
        <p>We utilized three medical external resources, TREC CDS, OHSUMED, and
ICD10, which were extracted from Wikipedia. Tables 1 show a summary of the TREC
CDS and OHSUMED collections. TREC CDS consists of biomedical literature,
specifically a subset of PubMed Central. A document is a full-text XML of a journal
article. OHSUMED consists of biomedical literature which is a subset of the
clinically oriented MEDLINE. Clearly,  = {  ℎ,   ,   } for the CBEEM.
Lucene3 was exploited to index and search the initial documents   . Stop-words were
removed using 419 stop-words4 in INQUERY. In addition, numbers were normalized
to NU&lt;# of DIGITS&gt;. A query-likelihood method with Dirichlet smoothing was chosen
3 http://lucene.apache.org/
4
http://sourceforge.net/p/lemur/galago/ci/default/tree/core/src/main/resources/stopwords/inquery
as a scoring function. |  | was set to 1000. Based on   , we performed eight runs
by differentiating the combining components of our re-ranking framework. Table 2
shows the descriptions and Tables 3 and 4 summarize the performances of the
submitted runs. The performances were measured by P@10, NDCG@10, rank-biased
precision (RBP), and two different variants of RBP (i.e., uRBP, and uRBPgr). In contrast to
the evaluation settings used in previous years, the readability of the retrieved medical
content, along with the common topical assessments of relevance, is added as new
evaluation measure [15].
3.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Results</title>
        <p>
          Table 2 describes our submitted runs for
          <xref ref-type="bibr" rid="ref6">CLEF 2015</xref>
          eHealth Task 2 and Table 3
summarizes our results obtained from the task’s official standard evaluation set. Runs 7 and
8 are different from runs 5 and 6, as the experiments were performed with expanded
queries produced from the CBEEM for ESA and CBDC, while runs 5 and 6 used
original queries.
        </p>
        <p>According to Table 3, ESA and CBDC using the concept relevant to ICD-10 are not
helpful according to a comparison of runs 1, 2 and 3. It can be concluded that the
reduction of the concept space without precise ICD-10 concepts resulted in low
discrimination power. On the other hand, the CBEEM showed consistent improvements over
QLD.</p>
        <p>The best performance was obtained in runs 6 and 8, where the CBEEM and the
CBDC were combined. This finding indicates that the use of external medical resources
when also considering concept-level associations can have synergetic effects on the
reranking of documents when they are in the proper right sequence. Moreover, the CBDC
is not apparently affected by the query expansion results.</p>
        <p>Run
1
2
3
4
5
6
7
8</p>
      </sec>
      <sec id="sec-3-3">
        <title>Description</title>
        <p>Query likelihood method with Dirichlet smoothing (QLD)
QLD + Explicit semantic analysis (ESA)
QLD + Concept-based document centrality (CBDC) using ESA
QLD + Cluster-based external expansion model (CBEEM)
QLD + CBEEM+ ESA
QLD + CBEEM+ CBDC
QLD + CBEEM + ESA with expanded query</p>
        <p>QLD + CBEEM + CBDC with expanded query</p>
        <p>In comparison with the readability-based measures (i.e., uRBP and uRBPgr), the best
results in RBP were obtained from runs 6 and 8. However, the best performances of the
two readability-based measures were observed from run 7.</p>
        <p>The results show that the selection of re-ranking components is important because
some of them can degrade previously achieved levels of moderated performance. In
addition, we can expect additional performance improvements by combining two
different re-ranking components if their application sequence is appropriate.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>This working note describes our efforts to find high-performance combinations of
different re-ranking components which utilize external medical resources. Among the
different runs we attempted, runs 6 and 8 (where our proposed CBEEM and CBDC were
used) showed the best performance in P@10, NDCG@10, and RBP. These results
imply that the effective use of external medical resources for re-ranking can overcome the
innate limitations of naïve queries by laypeople. As our future work, to enhance the
proposed re-ranking components, we plan systematically to analyze symptom-wise
evidence residing in promising external medical resources.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Abdul-Jaleel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          et al.:
          <source>UMass at TREC</source>
          <year>2004</year>
          :
          <article-title>Novelty and HARD</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <source>Proceedings of Text REtrieval Conference (TREC)</source>
          .
          <article-title>(</article-title>
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          et al.:
          <article-title>Semantic concept-enriched dependence model for medical information retrieval</article-title>
          .
          <source>Journal of biomedical informatics. 47</source>
          ,
          <fpage>18</fpage>
          -
          <lpage>27</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Egozi</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          et al.:
          <article-title>Concept-Based Information Retrieval Using Explicit Semantic Analysis</article-title>
          .
          <source>ACM Transactions on Information Systems</source>
          .
          <volume>29</volume>
          ,
          <issue>2</issue>
          ,
          <fpage>1</fpage>
          -
          <lpage>34</lpage>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          et al.:
          <source>Overview of the CLEF eHealth Evaluation Lab</source>
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>CLEF 2015 - 6th Conference and Labs of the Evaluation Forum. Lecture Notes in Computer Science (LNCS)</source>
          , Springer (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Hersh</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          et al.:
          <article-title>OHSUMED: an interactive retrieval evaluation and new large test collection for research</article-title>
          .
          <source>Proceedings of the 17th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '94</source>
          . pp.
          <fpage>192</fpage>
          -
          <lpage>201</lpage>
          (
          <year>1994</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Kurland</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>PageRank without hyperlinks: Structural re-ranking using links induced by language models</article-title>
          .
          <source>Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '05</source>
          . pp.
          <fpage>306</fpage>
          -
          <lpage>313</lpage>
          ACM Press, New York, New York, USA (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Lafferty</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhai</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Document language models, query models, and risk minimization for information retrieval</article-title>
          .
          <source>Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '01</source>
          . pp.
          <fpage>111</fpage>
          -
          <lpage>119</lpage>
          ACM Press, New York, New York, USA (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>Proceedings of the 27th annual international conference on Research and development in information retrieval - SIGIR '04</source>
          . pp.
          <fpage>186</fpage>
          -
          <lpage>193</lpage>
          ACM Press, New York, New York, USA (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Oh</surname>
            ,
            <given-names>H.-S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Myaeng</surname>
          </string-name>
          , S.-H.:
          <article-title>Utilizing global and path information with language modelling for hierarchical text classification</article-title>
          .
          <source>Journal of Information Science</source>
          .
          <volume>40</volume>
          ,
          <issue>2</issue>
          ,
          <fpage>127</fpage>
          -
          <lpage>145</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Palotti</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          et al.:
          <source>CLEF eHealth Evaluation Lab</source>
          <year>2015</year>
          ,
          <article-title>task 2: Retrieving Information about Medical Symptoms</article-title>
          .
          <source>CLEF 2015 Online Working Notes.</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Simpson</surname>
            ,
            <given-names>M.S.</given-names>
          </string-name>
          et al.:
          <article-title>Overview of the TREC 2014 Clinical Decision Support Track</article-title>
          .
          <source>Proceedings of Text REtri eval Conference (TREC)</source>
          .
          <article-title>(</article-title>
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          et al.:
          <article-title>A Study of Concept-based Weighting Regularization for Medical Records Search</article-title>
          .
          <source>Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics</source>
          . pp.
          <fpage>603</fpage>
          -
          <lpage>612</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>ACM Transactions on the Web. 6</source>
          ,
          <issue>4</issue>
          ,
          <fpage>1</fpage>
          -
          <lpage>29</lpage>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Zuccon</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koopman</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Integrating Understandability in the Evaluation of Consumer Health Search Engines</article-title>
          .
          <source>Proceedings of the SIGIR Workshop on Medical Information Retrieval (MedIR)</source>
          . (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>