<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>JU_KS_Group@FIRE 2016: Consumer Health Information Search</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kamal Sarkar</string-name>
          <email>jukamal2001@yahoo.com</email>
          <email>jukamal2001@yahoo.com Indra Banerjee Dept. of Computer Sc. &amp; Engg. Jadavpur University Kolkata, WB 700032 ardnibanerjee@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Debanjan Das</string-name>
          <email>dasdebanjan624@gmail.com</email>
          <email>dasdebanjan624@gmail.com Mamta Kumari Prasenjit Biswas Dept. of Computer Sc. &amp; Engg. Dept. of Computer Sc. &amp; Engg. Jadavpur University Jadavpur University Kolkata, WB 700032 Kolkata, WB 700032 mamta.mk222@gmail.com p.biswas.ju94@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. of Computer Sc. &amp; Engg., Jadavpur University</institution>
          ,
          <addr-line>Kolkata, WB 700032</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we describe the methodology used and the results obtained by us for completing the tasks given under the shared task on Consumer Health Information Search (CHIS) collocated with the Forum for Information Retrieval Evaluation (FIRE) 2016, ISI Kolkata. The shared task consists of two sub-tasks - (1) task1: given a query and a document/set of documents associated with that query, the task is to classify the sentences in the document as relevant to the query or not and (2) task 2: the relevant sentences need to be further classified as supporting the claim made in the query, or opposing the claim made in the query. We have participated in both the sub-tasks. The percentage accuracy obtained by our developed system for task1 was 73.39 which is third highest among the 9 teams participated in the shared task.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Consumer health information search</kwd>
        <kwd>searching behavior</kwd>
        <kwd>search tasks</kwd>
        <kwd>user query</kwd>
        <kwd>document sentences</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Categories and Subject Descriptors</title>
      <p>H.1.2 [Information Systems]: User/Machine Systems – human
factors, human information processing.</p>
    </sec>
    <sec id="sec-2">
      <title>1. INTRODUCTION</title>
    </sec>
    <sec id="sec-3">
      <title>1.1 Our Motivation</title>
      <p>
        A large number of websites provide health related information
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Consumer use of the Internet for seeking health
information is rapidly growing [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. By 1997, nearly half of
Internet users in the US had sought health information [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Expressed in raw numbers, an estimated 18 million adults in the
US sought health information online in 1998. The majority of
consumers seek for themselves health information related to
diseases for consultation with their physicians [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Information
found trough search on the web may influence medical decision
making and help consumers to manage their own care [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The
most common topics which are searched on the web are the
leading causes of death (heart disease and cancer) and Children
health.
      </p>
      <p>
        Information access mechanisms for factual health information
retrieval have matured considerably, with search engines
providing Fact Checked Health Knowledge Graph search results
to factual health queries. It is pretty straightforward to get an
answer to the query “what are the symptoms of Diabetes” from
these search engines [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ][
        <xref ref-type="bibr" rid="ref9">9</xref>
        ][
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. But the most general purpose
search engines can hardly find the answers of the complex health
search queries which do not have a single definitive answer and
whose answers have multiple perspectives. There may have a
search queries for which there are a large number of search results
reflecting the different perspectives and view-points in favor or
against the query.
      </p>
      <p>The term “Consumer Health Information Search” (CHIS) has
been used by the organizers of the shared task on Consumer
Health Information Search @FIRE 2016 to denote such
information retrieval search tasks for which there are no “Single
Correct Answer(s)” and instead, multiple and diverse
perspectives/points of view, which very often are contradictory in
nature, are available on the web regarding the queried
information1.</p>
    </sec>
    <sec id="sec-4">
      <title>1.2 Problem Statement</title>
      <p>The shared task on Consumer Health Information Search @FIRE
2016 has the following two sub-tasks:
A) Task 1- Given a CHIS query and a document/set of documents
associated with that query, the task given was to classify the
sentences in the document as relevant to the query or not.
Relevant sentences in the document being those which are useful
in providing the answer to the query.</p>
      <p>B) Task 2- These relevant sentences had to be further classified as
supporting the claim made in the query or opposing it.
1.2.1 Examples
E.g. Query - Are e-cigarettes safer than normal cigarettes?
S1: Because some research has suggested that the levels of most
toxicants in vapor are lower than the levels in smoke, e-cigarettes
have been deemed to be safer than regular cigarettes.
1 https://sites.google.com/site/multiperspectivehealthqa/
S2: David Peyton, a chemistry professor at Portland State
University who helped conduct the research, says that the type of
formaldehyde generated by e-cigarettes could increase the
likelihood it would get deposited in the lung, leading to lung
cancer.</p>
      <sec id="sec-4-1">
        <title>A)Relevant, B) oppose S3: Harvey Simon, MD, Harvard Health Editor, expressed concern that the nicotine amounts in e-cigarettes can vary significantly.</title>
      </sec>
      <sec id="sec-4-2">
        <title>A) Irrelevant, B) Neutral</title>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>2. METHODOLOGY</title>
    </sec>
    <sec id="sec-6">
      <title>2.1 Description</title>
      <p>For both the tasks-Task 1 and Task 2, we have used support
vector machines (SVM) as the classifier, but the feature sets for
the task1 and task2 were different. We discuss the feature sets
used for task1 and task 2 in sub section 2.1.1 and sub-section
2.1.3 respectively.</p>
      <sec id="sec-6-1">
        <title>2.1.1 Our Used Features for Task1</title>
        <p>For the task 1, we were given by the organizers of the shared task
a set of excel files where the heading of each excel file was a user
query. Each excel file contained a set of sentences that were
labeled as relevant or not relevant to the user query. The
sentences in these given training excel files were already labeled
as ‘relevant’ or ‘irrelevant’. We took each and every sentence
from each excel file and pair it with the corresponding query,
examined them, and calculated a set of five features discussed in
this sub section.
2.1.1.1 Exact Matching: We matched each sentences with
the user given query, word by word, and calculated the similarity
between the user query and the current sentence in the excel file;
e.g.</p>
        <p>Let the user query be “Ram is a good boy” and the current
sentence be “Shyam is a bad boy”. Between the user query and
the current sentence there are three words which are exactly
matching, i.e. “is”, “a” &amp; “boy”. Now the similarity between
these two strings is given as;
Similarity = {2 * (No. of Common Words)} / {(No. of words in
user query) + (No. Of words in the current sentence)} --- (i)
where, no. of Common Words = Number of words common to
both the user query and the current sentence.
2.1.1.2 Stemmed Word Matching: We stemmed both the
user query and the current sentence using a stemming tool
available in Python programming language. Stemming normalizes
a word by cutting out the excess part of a word due to
pluralization, or if the word is an adverb; e.g. mangoes → mango,
highly → high etc. After stemming we again calculated the
similarity between both the strings using equation (i).
2.1.1.3 Noun matching: We found, on a perusal of initial
sample data, that the nouns present in each sentence largely
influenced whether a search result was relevant or irrelevant to
the user query. So we isolated the nouns present in the user query,
searched whether any of these nouns were matching with any
word present in current sentence, and by this process we found
out the number of nouns present in the current sentence that were
exactly matching the nouns present in the user query. We
calculated the noun matching similarity using the following
formula;
Noun Similarity = (No. of nouns that are exactly matching with
the nouns in the query) / (No. of nouns present in the user query)
--- (ii)</p>
      </sec>
      <sec id="sec-6-2">
        <title>2.1.1.4 Neighborhood Matching: There were some words</title>
        <p>present in the sentences which were not matching exactly with the
words of user query, but they are semantically similar with the
user query words; e.g.</p>
        <p>
          Let ‘skin cancer’ be present in the user query and ‘melanoma’ be
present in the current sentence. Both words are spelt differently
but their meanings are similar, i.e. they are meaningfully similar.
To check whether the words were equivalent or not, we took each
word from the current sentence, searched it in our self-made
Wikipedia Dictionary [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], and extracted the first three sentences
describing that word’s meaning. We then matched the user query
words with the words present in the extracted sentences, and if the
word is present, we consider it as a match and, finally we
calculate the similarity again between the user query and the
current sentence using equation (i).
        </p>
        <p>We create the Wikipedia Dictionary by saving words along with
their meanings, which were extracted from Wikipedia. We use
our developed python script for creating this dictionary.</p>
      </sec>
      <sec id="sec-6-3">
        <title>2.1.1.5 COSINE Similarity:</title>
        <p>We represent both the query and a sentence using bag-of-words
model and each query as well as the sentence is represented as
vector. The component of each vector is TFIDF weight of a word
t which is calculated as follows:
IDF (t) = log(N/DF)
Where N= Total number of sentences and DF= Number of
sentences with word ‘t’ in it
TF(t) = (Number of times word ‘t’ appears in a sentence) / (Total
number of words in the sentence)
After calculating the vectors for the query and the sentence, the
cosine similarity between the query vector and the sentence
vector is calculated. The cosine similarity value is used as one of
feature values for relevance checking.</p>
      </sec>
      <sec id="sec-6-4">
        <title>2.1.2 Search as Classification</title>
        <p>For task 1, we represent each training sentence as vector of five
feature values mentioned above and label each vector as
“relevant” or “not relevant”. With this labeled training data, we
train the support vector machines (SVM). For SVM, we have used
SVC tool available in Python scikit learn and a model is
generated. Since no development set was available, for parameter
tuning, we split the training data into two parts-(1) the first part
contains 60% of the training data and second part contain 40% of
the training data. We train SVM with the 60% of the training data
and then we test the obtained model on the remaining part of the
training data. Thus we tune the parameters to obtain the best
parameter settings. Finally, we obtain the best results with the
settings where the cost parameter C set to 107, gamma set to 0.006
and kernel set to “poly”.</p>
        <p>Like training data, we represent the unlabeled test data released
by the organizers of the shared task in the similar way using the
five features mentioned in sub-section 2.1.1, and then submit it to
the trained classifier. The classifier, using its knowledge from
previous training data, predicts the labels for each of the sentences
present in test data.</p>
      </sec>
      <sec id="sec-6-5">
        <title>2.1.3 Our Used Features for Task2</title>
        <p>After relevancy checking (Task 1), Task 2 is carried out. By task
1, all the sentences in the excel file are divided into two classes;
(a) relevant and (b) irrelevant. Now the task is to determine
whether a relevant sentence was supporting the user query,
opposing the user query, or neutral with regard to the user query.
For this task we again calculated a set of N+4 features, where N =
number of distinct words present in the entire training files. Here
the feature set includes N number of distinct unigrams present in
the training data and four other features discussed in the following
sub-sections.
2.1.3.1 Number of Positive Words: We calculated the
number of positive words that were present in each sentence of
the excel file. We recognized the positive words from a particular
sentence by using a Python package called SentiWordNet2.
2.1.3.2 Number of Negative Words: We calculated the
number of negative words that were present in each sentence of
the excel file. We recognized the negative words from a particular
sentence by using a Python package called SentiWordNet.
2.1.3.3 Number of Neutral Words: We had already found
out the positive and negative words for a particular sentence, so
the words that were neither negative nor positive were classified
as neutral words and their occurrence in the current sentence was
counted.
2.1.3.4 Relevant or Irrelevant: In Task-1 we have already
labeled each sentence to be either relevant or irrelevant. We took
this label into consideration for this task. This was a binary
feature as the current sentence could either be relevant or
irrelevant.
2.1.3.5 ‘N’ Features: we represent each sentence as a
bag-ofwords model. According to vector space model, a sentence is
represented as N-dimensional vectors where N is the distinct
number of unigrams present in the training data. Weight of a word
used as the component of a vector is calculated using TFIDF
formula.
2.1.4 Sentiment Classification
We represent each sentence in the excel file as a vector using the
above mentioned N+4 features and label each vector with the
label of the corresponding training sentence. The label can be one
of three types- “Support”, “Oppose” and “Neutral”. Finally, we
submit labeled vectors to the SVM classifier as specified in the
Task-1 and trained it using them. The model is generated after
training. Like the task 1, we also we split the training data into
two parts-(1) the first part contains 60% of the training data,
which is used to develop the initial model and (2) the remaining
40% of the training data is used to test the model while tuning the
parameters. After tuning the parameters of SVC tool available in
Python scikit learn, we obtain the best model with the cost
2 http://www.nltk.org/howto/sentiwordnet.html
http://www.nltk.org/_modules/nltk/corpus/reader/sen
tiwordnet.html
parameter C set to 107, gamma set to 0.005 and kernel set to
“rbf”.</p>
        <p>We also represent unlabeled test data released by the organizers
for the task 2 as the vectors using the same feature set consisting
of N+4 features and submit them to the trained model which in
turn predicts label ‘supporting’/’opposing’/’neutral’ for each
sentence present in the test excel file.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>2.2 Architecture</title>
      <p>The architecture of our developed system used for task 1 and task
2 are shown in Figure 1 and Figure 2 respectively. For both the
systems, the important modules are feature extraction and
classifier. For the task 1, we have 5 features discussed in the
earlier sections and for task 2, we have used N + 4 features which
are also discussed in the earlier sections.</p>
      <p>For task 1, after feature extraction from each query-sentence
pairs, each sentence is represented as a vector which is labeled
with the label of the corresponding training sentence. Then the
labeled vectors are given to the classifier to produce a model.
Finally the learned model is used to determine the relevancy of
the test sentences given a query. For task 2, we extract features
from the sentences and sentences are represented as the vectors
labeled with one of the categories-“oppose”, “support” and
“neutral”. The classifier is trained with the labeled training pattern
vectors and the learned model is used to classify the test sentences
into one of categories-“oppose”, “support” and “neutral”.</p>
    </sec>
    <sec id="sec-8">
      <title>3. DATA SETS, RESULTS, EVALUATION</title>
    </sec>
    <sec id="sec-9">
      <title>3.1 Data Sets</title>
      <p>
        For the training data, we were given five user queries along with
number of sentences per query [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
• “does_sun_exposure_cause_skin_cancer”
-- 68 sentences
-- 83 sentences
-- 61 sentences
-- 71 sentences
-- 65 sentences
-- 342 sentences
-- 414 sentences
-- 260 sentences
-- 279 sentences
-- 247 sentences
• “e – cigarettes”
• “HRT_cause_cancer”
• “MMR_vaccine_lead_to_autism”
• “vitamin_C_common_cold”
A total of 348 sentences were present in the training data set.
For the test data, the queries were the same as the training data
and the number of unlabeled sentences per query given was as
follows.
• “does_sun_exposure_cause_skin_cancer”
• “e – cigarettes”
• “HRT_cause_cancer”
• “MMR_vaccine_lead_to_autism”
• “vitamin_C_common_cold”
A total 1542 sentences were present in the test data set
      </p>
    </sec>
    <sec id="sec-10">
      <title>3.2 Results</title>
      <p>
        We developed our systems for both task 1 and task 2 using the
training data [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] supplied to us by the organizers of the contest.
      </p>
      <sec id="sec-10-1">
        <title>User Query Training Data Wikipedia Dictionary</title>
        <p>Feature Extraction</p>
      </sec>
      <sec id="sec-10-2">
        <title>User Query Test Data</title>
        <p>Test Data Representation</p>
      </sec>
      <sec id="sec-10-3">
        <title>Sentences … F1 …</title>
        <p>F2
…</p>
        <p>F3
…</p>
        <p>F4
…</p>
        <p>F5
…
Training Data Representation</p>
      </sec>
      <sec id="sec-10-4">
        <title>Sentences … F1 …</title>
        <p>Test Data Representation
After release of test data by the organizers, we run our system on
the test data and send the result files along with the complete
system to the organizers. They evaluated the results using the
traditional percentage accuracy and published the results which
were sent to us through e-mail.</p>
        <p>We have shown the officially published results of task 1 and task
2 for the 9 participating teams in Table 1 and Table 2
respectively. The results shown in red bold font are the
performances of top systems participated in the tasks.
Out of the 9 participants, our system (JU_KS_Group) achieves the
third highest average accuracy for task 1, i.e. 73.39257557%. We
can evaluate the results for task 1 in a different angle. It is evident
from Table 1 that our system performs better for 3 queries out of
5 queries whereas the system SSN_NLP with the best average
accuracy (78.10%) performs better for 2 queries out of five
queries. The main reason for my system giving better results for
task 1 is the use of two novel features, noun matching and
neighborhood matching.</p>
        <p>For the task 2, our system achieves an average accuracy of
33.63514278%. For the task 2, our system achieves relatively
poor performance. One of the reasons of getting poor performance
for task 2 is that we have considered “neutral” class along with
other two classes “oppose” and “support” while classifying the
relevant sentences. It is evident from the training data that only
the irrelevant sentences in the training data were assigned the
“neutral” class. Actually the task2 was to classify the relevant
sentences into two categories-“Support” and “oppose”, but we
have mistakenly considered the task2 as 3-class problem instead
of 2-class problem. We are working to improve our proposed
methods so that our systems can perform more accurately for both
the tasks.</p>
      </sec>
    </sec>
    <sec id="sec-11">
      <title>4. CONCLUSION</title>
      <p>There has been a dearth of proper searching systems for medical
queries and our work on the CHIS tasks put us on the path to
filling this void. The methodology we used can be improved on
and innovated with to create a novel searching method for not
only medical queries, but any specific search queries of any field.
What we have done, and our continuing to improve on, is a
logical way of searching through data which is already available
to the public. We sincerely believe that through machine learning
and natural language processing, the future of online searching
can be achieved; and have tried to contribute towards this goal
through our paper. And that this will especially be of use in the
medical field.</p>
      <p>For future work, we would incorporate a word sense
disambiguation module to disambiguate the query words. We
hope that our system will give more accurate results for task 2 if
we consider classification of relevant sentences as 2-class
problem (“support” and “oppose”) instead of considering it as the
3-class (“support”, “oppose” and “neutral”) problem that we did
during the contest.</p>
      <sec id="sec-11-1">
        <title>ACKNOWLEDGMENTS We would like to thank the Forum for Information Retrieval Evaluation (FIRE) 2016, ISI Kolkata, for providing us the tasks and datasets for C.H.I.S.</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]. Cline,
          <string-name>
            <given-names>R. J.</given-names>
            , &amp;
            <surname>Haynes</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M.</surname>
          </string-name>
          (
          <year>2001</year>
          ).
          <article-title>Consumer health information seeking on the Internet: the state of the art</article-title>
          .
          <source>Health education research</source>
          ,
          <volume>16</volume>
          (
          <issue>6</issue>
          ),
          <fpage>671</fpage>
          -
          <lpage>692</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>.</given-names>
            <surname>Grandinetti</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. A.</surname>
          </string-name>
          (
          <year>2000</year>
          )
          <article-title>Doctors and the Web: help your patients surf the Net safely</article-title>
          .
          <source>Medical Economics</source>
          , April,
          <fpage>28</fpage>
          -
          <lpage>34</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>.</given-names>
            <surname>Lacroix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            ,
            <surname>Backus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            and
            <surname>Lyon</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. J.</surname>
          </string-name>
          (
          <year>1994</year>
          )
          <article-title>Service providers and users discover the Internet</article-title>
          .
          <source>Bulletin of the Medical Library Association</source>
          ,
          <volume>82</volume>
          ,
          <fpage>412</fpage>
          -
          <lpage>418</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]. Eng,
          <string-name>
            <given-names>T. R.</given-names>
            ,
            <surname>Maxfield</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Patrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Deering</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            ,
            <surname>Ratzan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            and
            <surname>Gustafson</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. H.</surname>
          </string-name>
          (
          <year>1998</year>
          )
          <article-title>Access to health information and support: a public highway or a private road?</article-title>
          <source>Journal of the American Medical Association</source>
          ,
          <volume>280</volume>
          ,
          <fpage>1371</fpage>
          -
          <lpage>1375</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>.</given-names>
            <surname>Chi-Lum</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          (
          <year>1999</year>
          )
          <article-title>Friend or foe: consumers using the Internet for medical information</article-title>
          .
          <source>Journal of Medical Practice Management</source>
          ,
          <volume>14</volume>
          ,
          <fpage>196</fpage>
          -
          <lpage>198</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]. Boyer,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Selby</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            and
            <surname>Appel</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. D.</surname>
          </string-name>
          (
          <year>1998</year>
          )
          <article-title>The Health on the Net Code of Conduct for medical and health Web sites</article-title>
          .
          <source>Medinfo, 9 (Part 2)</source>
          ,
          <fpage>1163</fpage>
          -
          <lpage>1166</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>.</given-names>
            <surname>Wilkins</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. S.</surname>
          </string-name>
          (
          <year>1999</year>
          )
          <article-title>Expanding Internet access for health care consumers</article-title>
          .
          <source>Health Care Management Review</source>
          ,
          <volume>24</volume>
          ,
          <fpage>30</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]. Hong,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Cruz</surname>
          </string-name>
          . N.,
          <string-name>
            <surname>Marnas</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Early</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Gillis</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>2002</year>
          .
          <article-title>A query analysis of consumer health information retrieval</article-title>
          .
          <source>In the Proceedings of AMIA</source>
          <year>2002</year>
          ,
          <volume>1046</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]. Belkin,
          <string-name>
            <surname>N.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Oddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. N.</given-names>
            , and
            <surname>Brooks</surname>
          </string-name>
          ,
          <string-name>
            <surname>H. M.</surname>
          </string-name>
          <year>1982</year>
          .
          <article-title>Ask for information retrieval: Part I. background and theory</article-title>
          .
          <source>Journal of Documentation</source>
          ,
          <volume>38</volume>
          ,
          <fpage>61</fpage>
          -
          <lpage>71</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]. Keselman,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Browne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.C.</given-names>
            , and
            <surname>Kaufman</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.R.</surname>
          </string-name>
          <year>2008</year>
          .
          <article-title>Consumer health information seeking as hypothesis testing</article-title>
          .
          <source>JAMIA</source>
          ,
          <volume>15</volume>
          ,
          <fpage>484</fpage>
          -
          <lpage>495</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]. Efthimiadis,
          <string-name>
            <surname>E.N.</surname>
          </string-name>
          <year>2009</year>
          .
          <article-title>How students search for consumer health information on the web</article-title>
          .
          <source>In Proceedings of the 42nd Hawaii International Conference on System Sciences</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]. Sinha,
          <string-name>
            <given-names>M.</given-names>
            and
            <surname>Mannarswamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            and
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>CHIS@FIRE: Overview of the CHIS Track on Consumer Health Information Search</article-title>
          , Working notes of FIRE 2016 -
          <article-title>Forum for Information Retrieval Evaluation, Kolkata</article-title>
          , India, December 7-
          <issue>10</issue>
          ,
          <year>2016</year>
          , CEUR Workshop Proceedings, CEUR-WS.org.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>