<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ECSTRA-INSERM @ CLEF eHealth2016-task 2: ICD10 Code Extraction from Death Certi cates</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mohamed Dermouche</string-name>
          <email>mohamed.dermouche@inserm.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vincent Looten</string-name>
          <email>vincent.looten@aphp.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>R´emi Flicoteaux</string-name>
          <email>remi.flicoteaux@aphp.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sylvie Chevret</string-name>
          <email>sylvie.chevret@univ-paris-diderot.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julien Velcin</string-name>
          <email>julien.velcin@univ-lyon2.fr</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Namik Taright</string-name>
          <email>namik.taright@aphp.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AP-HP</institution>
          ,
          <addr-line>Paris, F-75004</addr-line>
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>INSERM, U1153 Epidemiology and Biostatistics Sorbonne Paris Cit ́e Research Center (CRESS), ECSTRA team</institution>
          ,
          <addr-line>Paris, F-75010</addr-line>
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Paris Diderot University</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Saint-Louis Hospital, AP-HP</institution>
          ,
          <addr-line>Paris, F-75010</addr-line>
          <country country="FR">France</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Universit ́e de Lyon (ERIC Lyon 2)</institution>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes the participation of ECSTRA-INSERM team at CLEF eHealth 2016, task 2.C. The task involves extracting ICD10 codes from death certificates, mainly described with short plain texts. We cast the task as a machine learning problem involving the prediction of the ICD10 codes (categorical variable) from the raw text transformed into a bag-of-words matrix. We rely on probabilistic topic models that we evaluate against classical classifiers such as SVM and Naive Bayes. We demonstrate the effectiveness of topic models for this task in terms of prediction accuracy and result interpretation.</p>
      </abstract>
      <kwd-group>
        <kwd>ICD10 code assignment</kwd>
        <kwd>cause of death extraction</kwd>
        <kwd>topic models</kwd>
        <kwd>machine learning</kwd>
        <kwd>natural language processing</kwd>
        <kwd>text mining</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Completing death certificates is a routine task in hospitals and healthcare
institutions. In France, the death certificates are produced by physicians and
transmitted to the French Epidemiological Center for the Causes of Death (C´epiDC)6.
Beyond the administrative and personal information, the death certificates
usually contain a free-text description of the cause(s) of death. To monitor
population’s health, free texts are converted by the C´epiDC into formal standardized
codes, usually derived from the International Classification of Diseases (ICD)
taxonomy. These codes also serve as a basis for mortality and epidemiology
studies.</p>
      <p>The ICD taxonomy covers a wide range of diseases, symptoms, signs,
procedures, and other content related to diseases7. The World Health Organization</p>
    </sec>
    <sec id="sec-2">
      <title>6 http://www.cepidc.inserm.fr/ 7 http://www.who.int/classifications/icd/en/</title>
      <p>issues separate versions of ICD per language/country. In this paper, we use the
French release of ICD, which is now at its 10th revision (called ICD10). It covers
more than 20,000 codes including diagnoses and procedures, but only a subset
of theses codes can be causes of death. An example is provided in Table 1.</p>
      <p>
        Requiring manual work and expertise, the task of ICD10 code extraction
from text is quite time-consuming because the ICD10 taxonomy contains
thousands of possible causes of death. Within the CLEF eHealth 2016, the task 2.C
focuses on the problem of automatic extraction of the causes of death from the
textual description [
        <xref ref-type="bibr" rid="ref10 ref8">8, 10</xref>
        ]. The task may be approached either from a machine
learning perspective (supervised classification) or a natural language processing
perspective by using syntactic and/or semantic decision rules. Both approaches
aim at automating the ICD10 code extraction from death certificates.
      </p>
      <p>
        In this paper, we describe our system to automatic cause-of-death extraction
following the first approach. Concerning the used methods, we mainly focus on
probabilistic topic models [
        <xref ref-type="bibr" rid="ref14 ref3">3, 14</xref>
        ] that we evaluate against traditional machine
learning methods, like SVM and Naive Bayes, with respect to predictive
accuracy and result interpretation. We show that topic models are competitive with
traditional methods in terms of predictive accuracy. We also show that topic
models offer more easily-interpreted results and allow to gain a better insight in
the data analysis.
2
      </p>
      <sec id="sec-2-1">
        <title>Methods</title>
        <p>
          Topic models are probabilistic approaches to discovering hidden structures
(commonly called topics) from text. Topic models have shown significant efficiency
over baseline models for various language modeling and text mining tasks, like
topic discovery [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], information retrieval [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], and text classification [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. The
general approach is to model the co-occurrence relation among words (observed
variables) and build soft clusters of words characterizing the topics.
        </p>
        <p>
          Latent Dirichlet Allocation (LDA) [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] is one of the most popular topic models
completely built on the co-occurrence assumption. In LDA, the words that tend
to co-occur in the same documents are more likely to characterize the same topic.
Conversely, the words that rarely co-occur are likely to describe different topics.
In LDA, the document’s words are captured using an observed variable w (see
Figure 1.a) while the topics are inferred as a latent variable z using Bayes
calculus. For inferring the latent variables, a number of optimization methods can be
deployed, like Gibbs sampling, Expectation Maximization, etc. For more details
on the optimization process, we refer the uninitiated reader to the tutorials in [
          <xref ref-type="bibr" rid="ref15 ref6">6,
15</xref>
          ].
        </p>
        <p>
          Relying solely on word co-occurrence, LDA model is fully unsupervised. In
this work, we rely on a supervised extension of LDA called LabeledLDA [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
Unlike LDA, LabeledLDA is supervised in that it learns, in addition to the
hidden topics, a “response variable”. For many tasks, LabeledLDA has shown
competitive results compared to traditional machine learning models, like Naive
Bayes and SVM.
        </p>
        <p>To better explain how LabeledLDA works, we rely on the plate notation given
in Figure 1.b. In the traditional machine learning methods, the words (variable
w) directly influence the prediction decision. For example, in Naive Bayes
classifier, the decision depends on the word frequencies per class. In LabeledLDA,
The prediction decision depends on both document’s words and topics. That is,
the topic’s proportions of the document to be classified are taken into account
when computing the document-code probabilities. On the other hand, the topic
extraction with LabeledLDA is influenced by the document’s classes. Thus, the
documents from the same class are likely to be linked to the same topics. This
feature allows LabeledLDA to take advantage of the hidden structures of
documents as well as the topic-code relations. For example, the code N189 (chronic
kidney disease) is likely to describe a topic about loss in kidney function, that
in turn can be described using the words “kidney”, “renal”, “failure”, etc.</p>
        <p>
          The efficiency of LabeledLDA for text classification has been proved in many
tasks, including diagnosis code assignment to medical summaries [
          <xref ref-type="bibr" rid="ref11 ref4">4, 11</xref>
          ], which
is very close to our task. In this work, we aim at experimenting LabeledLDA
model for the specific task of death cause extraction. The main differences from
the work in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] are the following:
{ The task is not the same, here the ICD10 codes correspond to the causes of
death while in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] the codes correspond to diagnoses.
{ The number of possible codes is much more important (3,231 vs. 60 in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ])
which endows the task with more challenging analysis problems.
{ The documents used here are much shorter (a document contains 3.6 words
on average vs. 60 words in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]).
        </p>
        <p>The parameters of LabeledLDA are fixed empirically in such a way to
maximize the predictive scores on a held-out dataset (20% documents randomly
sampled from the whole dataset). As such, the number of topics K is equal to
the number of codes (3,231), = 0:005; = 0:07. The remaining
hyperparameters are learnt from data. In addition, we perform two experiments by setting
the number of iterations to 20,000 and 50,000 respectively.</p>
        <p>The two other classifiers used in this work are SVM and Naive Bayes. SVM
is a non-probabilistic binary classifier that maps the documents in such a way to
maximize the gap between those from opposite classes. For our non-binary
problem, we use the one-vs-one technique that consists of learning a separate classifier
of each pair of classes then taking the class with the largest weight. Naive Bayes
is a simple and widely-used probabilistic classifier based on the calculation of
conditional probabilities (probability of words given the outcome class). The
process is then easily inverted using Bayes theorem and word independence
assumption. To run these methods, we rely on Python and “scikit-learn” package8,
specifically “BernoulliNB” and “LinearSVC” implementations. For SVM, we set
the penalty parameter c to 1.0. The rest of parameters are left to default values.
3</p>
      </sec>
      <sec id="sec-2-2">
        <title>Dataset</title>
        <p>
          The C´epiDC corpus has been created by the French Center for Epidemiology
and Medical Causes of Death (C´epiDC) specifically for the CLEF eHealth 2016
contest [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. It is composed of separate train/test samples of death certificates.
Only the textual description of the causes of death are available for analysis.
C´epiDC dataset is highly imbalanced: about 80% of documents are assigned to
less than 20% of codes.
        </p>
        <p>Also, in order to reduce dimensionality, we filter out frequent words
(contained in more than 50% of documents) and rare words (contained in less than
3 documents). We also remove stopwords and numerics. The preprocessed text
documents are then mapped into a bag-of-words representation where the words
are weighted according to their presence/absence in the document (binary
values). Table 2 gives an overview of the preprocessed C´epiDC sample used for
training.
4</p>
      </sec>
      <sec id="sec-2-3">
        <title>Results and Discussion</title>
        <p>The methods are evaluated and ranked based on micro-averaged F-score
(harmonic mean of precision and recall weighted by the class size). LabeledLDA</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>8 http://scikit-learn.org/</title>
      <p>model is put against other traditional machine learning methods: Naive Bayes
and SVM. We have chosen these methods among others because they gave the
best scores.</p>
      <p>The obtained results are given in Table 3. For the official evaluation, we
have submitted a run of LabeledLDA with 20,000 iterations because the running
time was too long. In Table 3, we also evaluate a run of LabeledLDA with 50,000
iterations. Our two official submissions are given in italic whereas the best scores
are given in bold.</p>
      <p>
        As can be seen, SVM achieves the best F-score compared to other
methods, followed by LabeledLDA model. In terms of micro-averaged F-score, SVM
achieves 75.19% while LabeledLDA arrives second with 73.53%. Compared to the
other 6 participant systems, our LabeledLDA-based system was ranked fourth
(based on 20,000 iterations). When rising the number of iterations to 50,000,
LabeledLDA would be ranked second. Based on the official results from [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], the
average score from all participants was 71.85% while the median was 69.97%.
      </p>
      <p>
        Although the superiority of SVM over LabeledLDA in terms of precision
score, LabeledLDA’s most interesting advantage resides in offering more
easilyinterpreted results. In Table 4, we show the top-10 words characterizing topics
extracted with LabeledLDA model. To complete these results, we have asked a
physician to decide of the significance of topics by looking at the top-10 words
globally. The physician has agreed that topics were extremely informative. Most
of the causes of death represented here could easily be recognized from the
associated topical words.
pleural (pleural), m´esoth´eliome (mesothelioma), malin
(malignant), droit (right), m´esoth´elium (mesothelium), gauche (left),
m´etastatique (metastatic), terminal (terminal), pl`evre (pleura),
´epithelioide (epithelioid)
sarcome (sarcoma), thoracique (thoracic), pulmonaire
(pulmonary), multim´etastatique (multi metastatic), art`ere (artery),
r´ecidive (recurrence), ´evolutif (evolutive), thorax (thorax),
pari´etal (parietal), paroi (lining)
insuffisance (failure), surr´enalienne (adrenal), aigu¨e (acute),
chronique (chronic), diab`ete (diabetes), surr´enale (adrenal),
r´enal (renal), insulino (insulin), requ´erant (petitioner),
hypophysaire (hypophyseal)
´epilepsie (epilepsy), AVC (stroke), ´epileptique (epileptic),
enc´ephalopathie (encephalopathy), alzheimer (alzheimer),
evolu´ee (evolved), syndrˆome (syndrome), s´equelles (aftermath),
post (after), maladie (disease)
The problem of ICD code extraction has been investigated from a larger
perspective involving code assignment to various types of medical documents. The cause
of death may be considered as a specific task. The majority of these works have
focused on English documents. Number of them have been published with the
Computational Medicine Center’s 2007 medical NLP contest involving ICD code
assignment to radiology reports [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The dataset contained 45 different diagnosis
codes and each document could be labeled with one or more codes (multi-label
task). [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] used BoosTexter, a boosting-like technique based on a weak classifier,
to learn a set of classification rules. In [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], a simple classifier was learnt based
on the presence/absence of medical concepts from UMLS ontology9. The best
micro-averaged F-score achieved within the contest was about 89%.
      </p>
      <p>
        Apart from this contest, in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] both SVM and Ridge Regression classifiers
have achieved a score of 68% on a dataset with 2,618 distinct codes and about
100,000 documents. In [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], SVM classifier has been tested considering both
flat and hierarchical setting. The hierarchical setting relies on the tree
structure of ICD codes to improve accuracy. On a dataset with 5,030 distinct codes,
the achieved F-scores were about 27% under flat setting and about 39% under
hierarchical setting.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], LabeledLDA has been used to extract ICD codes from both English and
French discharge summaries. Compared to SVM classifier, LabeledLDA achieved
almost same performance. With 60 distinct codes, the micro-averaged F-score
was about 52%. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the authors proposed a hierarchically-supervised topic
model (HSLDA) that combines traditional topic modeling with the ICD
struc9 https://www.nlm.nih.gov/research/umls/
ture. The hierarchical structure of ICD codes has been taken into account during
the topic learning step. Towards this end, the final predicted code was
constrained to derive from a single branch of the tree: a code could not be assigned
to the document if its parent in the tree were not. Compared to a non-hierarchical
version [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], HSLDA performed about 5% better on a dataset with 7,298 distinct
codes. The source code of HSLDA has not been made available, which prevented
from experimenting it for this task.
6
      </p>
      <sec id="sec-3-1">
        <title>Conclusion</title>
        <p>In this paper, we have described our system used to extract ICD10 codes from
death certificates within the CLEF 2016 eHealth, task 2.C. Our system is based
on LabeledLDA model: a supervised topic model for topic discovery and
document classification. Even if LabeledLDA does not outperform SVM classifier,
LabeledLDA provides an explanation of the results (why such code for such
document?) and allows a more in-depth understanding of the classification
mechanisms. This feature is obviously a clear advantage over machine learning
methods, like SVM that are usually based on complex mechanisms and consequently
less suitable for human interaction.</p>
        <p>
          As a promising future direction, we believe that the performance of
LabeledLDA can be improved by integrating an “active learning” component. That
is, the active learning allows capturing and integrating user’s feedback into the
learning process [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ]. As such, it will also be possible to focus on specific
data, for example a subset of misclassified documents, chosen by the user based
on her experience [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. The high flexibility offered by topic models will allow for
more in-depth error analysis and better understanding of data, which is more
convenient for integrating user interaction.
        </p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>David</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Blei</surname>
            and
            <given-names>Michael I.</given-names>
          </string-name>
          <string-name>
            <surname>Jordan</surname>
          </string-name>
          .
          <article-title>Modeling annotated data</article-title>
          .
          <source>In Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval (SIGIR'03)</source>
          , pages
          <fpage>127</fpage>
          -
          <lpage>134</lpage>
          , Toronto, Canada,
          <year>2003</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>David</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Blei</surname>
            and
            <given-names>Jon D.</given-names>
          </string-name>
          <string-name>
            <surname>Mcauliffe</surname>
          </string-name>
          .
          <article-title>Supervised topic models</article-title>
          .
          <source>In Advances in Neural Information Processing Systems (NIPS'07)</source>
          , pages
          <fpage>121</fpage>
          -
          <lpage>128</lpage>
          , Vancouver, Canada,
          <year>2007</year>
          . Curran Associates, Inc.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>David</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>Andrew Y.</given-names>
          </string-name>
          <string-name>
            <surname>Ng</surname>
            , and
            <given-names>Michael I. Jordan. Latent Dirichlet</given-names>
          </string-name>
          <string-name>
            <surname>Allocation</surname>
          </string-name>
          .
          <source>The Journal of Machine Learning Research (JMLR)</source>
          ,
          <volume>3</volume>
          :
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Mohamed</given-names>
            <surname>Dermouche</surname>
          </string-name>
          , Julien Velcin, R´emi Flicoteaux, Sylvie Chevret, and
          <string-name>
            <given-names>Namik</given-names>
            <surname>Taright</surname>
          </string-name>
          .
          <article-title>Supervised Topic Models for Diagnosis Code Assignment to Discharge Summaries</article-title>
          .
          <source>In Proceedings of the 17th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing'16)</source>
          , Konya, Turkey,
          <year>2016</year>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Ira</given-names>
            <surname>Goldstein</surname>
          </string-name>
          , Anna Arzrumtsyan, and
          <string-name>
            <given-names>Ozlem</given-names>
            <surname>Uzuner</surname>
          </string-name>
          .
          <article-title>Three approaches to automatic assignment of ICD-9-CM codes to radiology reports</article-title>
          .
          <source>In Proceedings of AMIA Symposium (AMIA'07)</source>
          , pages
          <fpage>279</fpage>
          -
          <lpage>83</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Gregor</given-names>
            <surname>Heinrich</surname>
          </string-name>
          .
          <article-title>Parameter estimation for text analysis</article-title>
          .
          <source>Technical report</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Yuening</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            Boyd-Graber,
            <given-names>Brianna</given-names>
          </string-name>
          <string-name>
            <surname>Satinoff</surname>
            , and
            <given-names>Alison</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <article-title>Interactive topic modeling</article-title>
          .
          <source>Machine Learning</source>
          ,
          <volume>95</volume>
          (
          <issue>3</issue>
          ):
          <fpage>423</fpage>
          -
          <lpage>469</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Liadh</given-names>
            <surname>Kelly</surname>
          </string-name>
          , Lorraine Goeuriot, Hanna Suominen,
          <article-title>Aur´elie N´ev´eol, Joao Palotti, and Guido Zuccon. Overview of the CLEF eHealth Evaluation Lab 2016</article-title>
          .
          <source>CLEF 2016 - 7th Conference and Labs of the Evaluation Forum. Lecture Notes in Computer Science (LNCS)</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Lucian</given-names>
            <surname>Vlad</surname>
          </string-name>
          <string-name>
            <surname>Lita</surname>
          </string-name>
          , Shipeng Yu,
          <string-name>
            <given-names>Stefan</given-names>
            <surname>Niculescu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Jinbo</given-names>
            <surname>Bi</surname>
          </string-name>
          .
          <article-title>Large Scale Diagnostic Code Classification for Medical Patient Records</article-title>
          .
          <source>In Proceeding sof the International Joint Conference on Natural Language Processing (IJCNLP'08)</source>
          , pages
          <fpage>877</fpage>
          -
          <lpage>882</lpage>
          , Hyderabad, India,
          <year>2008</year>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Aur</surname>
          </string-name>
          <article-title>´elie N´ev´eol, Lorraine Goeuriot</article-title>
          , Liadh Kelly, Kevin Cohen, Cyril Grouin, Thierry Hamon, Thomas Lavergne, Gr´egoire Rey, Aude Robert, Xavier Tannier, and
          <string-name>
            <given-names>Pierre</given-names>
            <surname>Zweigenbaum</surname>
          </string-name>
          .
          <article-title>Clinical information extraction at the CLEF eHealth Evaluation lab 2016</article-title>
          .
          <article-title>In CLEF 2016 Evaluation Labs</article-title>
          and Workshop: Online Working Notes, E´vora, Portugal,
          <year>2016</year>
          . CEUR-WS.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Adler</surname>
            <given-names>Perotte</given-names>
          </string-name>
          , Nicholas Bartlett,
          <string-name>
            <given-names>Frank</given-names>
            <surname>Wood</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Noemie</given-names>
            <surname>Elhadad</surname>
          </string-name>
          .
          <article-title>Hierarchically Supervised Latent Dirichlet Allocation</article-title>
          .
          <source>In Advances in Neural Information Processing Systems (NIPS'11)</source>
          , pages
          <fpage>2609</fpage>
          -
          <lpage>2617</lpage>
          , Granada, Spain,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Adler</surname>
            <given-names>Perotte</given-names>
          </string-name>
          , Rimma Pivovarov, Karthik Natarajan, Nicole Weiskopf, Frank Wood, and No´emie Elhadad.
          <article-title>Diagnosis code assignment: models and evaluation metrics</article-title>
          .
          <source>Journal of the American Medical Informatics Association (JAMIA)</source>
          ,
          <volume>21</volume>
          (
          <issue>2</issue>
          ):
          <fpage>231</fpage>
          -
          <lpage>237</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. John P. Pestian, Christopher Brew,
          <string-name>
            <given-names>Pawel</given-names>
            <surname>Matykiewicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Hovermale</surname>
          </string-name>
          , Neil Johnson, K. Bretonnel Cohen, and
          <string-name>
            <given-names>Wlodzislaw</given-names>
            <surname>Duch</surname>
          </string-name>
          .
          <string-name>
            <given-names>A Shared</given-names>
            <surname>Task</surname>
          </string-name>
          <article-title>Involving Multi-label Classification of Clinical Free Text</article-title>
          .
          <source>In Proceedings of the Workshop on BioNLP 2007: Biological, Translational, and Clinical Language Processing (BioNLP'07)</source>
          , pages
          <fpage>97</fpage>
          -
          <lpage>104</lpage>
          , Prague, Czech Republic,
          <year>2007</year>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Daniel</surname>
            <given-names>Ramage</given-names>
          </string-name>
          , David Hall,
          <string-name>
            <given-names>Ramesh</given-names>
            <surname>Nallapati</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Christopher D.</given-names>
            <surname>Manning</surname>
          </string-name>
          . Labeled LDA :
          <article-title>A supervised topic model for credit attribution in multi-labeled corpora</article-title>
          .
          <source>In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing (EMNLP'09)</source>
          ,
          <source>number August</source>
          , pages
          <fpage>248</fpage>
          -
          <lpage>256</lpage>
          , Singapore, Singapore,
          <year>2009</year>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>Philip</given-names>
            <surname>Resnik</surname>
          </string-name>
          and
          <string-name>
            <given-names>Eric</given-names>
            <surname>Hardisty</surname>
          </string-name>
          .
          <article-title>Gibbs Sampling for the Uninitiated</article-title>
          .
          <source>Technical Report June</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Yi</surname>
            <given-names>Yang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Doug</given-names>
            <surname>Downey</surname>
          </string-name>
          , and
          <string-name>
            <surname>Jordan</surname>
          </string-name>
          Boyd-graber.
          <article-title>Efficient Methods for Incorporating Knowledge into Topic Models</article-title>
          .
          <source>In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP'15)</source>
          , number r, pages
          <fpage>308</fpage>
          -
          <lpage>317</lpage>
          , Lisbon, Portugal,
          <year>2015</year>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Yi</surname>
            <given-names>Yang</given-names>
          </string-name>
          , Shimei Pan, Jie Lu, Mercan Topkara, and
          <string-name>
            <given-names>Doug</given-names>
            <surname>Downey</surname>
          </string-name>
          .
          <article-title>Incorporating User Input with Topic Modeling</article-title>
          .
          <source>In CIKM 2014 Workshop on Interactive Mining for Big Data (ImBig'14)</source>
          , Shanghai, China,
          <year>2014</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <given-names>Xing</given-names>
            <surname>Yi</surname>
          </string-name>
          and
          <string-name>
            <given-names>James</given-names>
            <surname>Allan</surname>
          </string-name>
          .
          <article-title>A Comparative Study of Utilizing Topic Models for Information Retrieval</article-title>
          .
          <source>In Proceedings of the 31th European Conference on IR Research on Advances in Information Retrieval (ECIR'09)</source>
          , pages
          <fpage>29</fpage>
          -
          <lpage>41</lpage>
          , Toulouse, France,
          <year>2009</year>
          . Springer-Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <given-names>Yitao</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <article-title>A hierarchical approach to encoding medical concepts for clinical notes</article-title>
          .
          <source>In Proceedings of the 46th Annual Meeting of the Association for Computational Linguistics on Human Language Technologies: Student Research Workshop (HLT-SRWS'08)</source>
          , pages
          <fpage>67</fpage>
          -
          <lpage>72</lpage>
          , Columbus,
          <string-name>
            <surname>OH</surname>
          </string-name>
          , USA,
          <year>2008</year>
          . ACL.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>