<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Medical Concept Resolution</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nitish Aggarwal</string-name>
          <email>nitish.aggarwal@insight-centre.org</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ken Barker</string-name>
          <email>kjbarker@us.ibm.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chris Welty</string-name>
          <email>chris.welty@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IBM Watson Research</institution>
          ,
          <addr-line>Yorktown Heights, NY</addr-line>
          ,
          <country>USA Insight</country>
          <institution>-Centre, National University of Ireland Galway</institution>
          ,
          <country country="IE">Ireland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we present a problem that we refer to as Medical Concept Resolution for nding concept identi ers in a large knowledge base, given medical terms mentioned in a text. We de ne the problem with its unique features and novel algorithms to address it. We compare performance to MetaMap and nd distinct and complementary behavior.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>vein". Relaxing the term order (e.g., \jugular vein distended") and substituting
synonyms (e.g., \engorged jugular vein"), we still nd no UMLS concepts with
matching labels. There are, however, concepts in UMLS for \jugular vein
distention" and \jugular venous distension".</p>
      <p>Multiplicity In concept repositories that combine multiple sources, there are
often multiple entries for the same domain concept. So even after a concept is
discovered, there may be other appropriate concepts. In UMLS there are many
CUIs with the label \pain". Therefore, MCR has to deal with concept
disambiguation similar to open domain entity linking.</p>
      <p>Granularity Even when there are close super cial matches between concept
labels and text spans, more speci c concepts often exist that capture more of
the semantics for the span, given a larger context. For instance, the CUI for
\jugular vein distention" may match a text span perfectly, but the more speci c
CUI for \jugular venous distension with inspiration" may be more appropriate
when considering a larger context.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Approach</title>
      <p>
        Our approach for mapping text spans (\mentions") to UMLS CUIs consists of
two main steps: candidate over-generation and candidate reranking. For
obtaining the mentions from text, we use a CRF-based method for extracting medical
terms [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
2.1
      </p>
      <sec id="sec-2-1">
        <title>Candidate Over-generation</title>
        <p>The intuition behind over-generation is that there may be a mismatch between
the mentions in text and the variant labels of target CUIs. Over-generation nds
all CUIs having any variant containing any of the tokens in the mention text.
The resulting candidates include many irrelevant CUIs, but also relevant CUIs
that are more general than the mention and CUIs that are more speci c. For
example, candidates for the string \pupil miosis" include the CUIs for pupil,
miosis, school pupil, congenital miosis, pupillary miosis of the right eye, etc.
Candidate over-generation may produce an enormous number of candidates.
For e ciency, only those candidates that are most similar to the mention are
considered in the subsequent reranking step. The most similar candidates are
determined by inverse document frequency (IDF) rank weighted similarity of
their labels to the mention text. The n tokens in the original mention are ranked
according to their IDF in a medical corpus. The ranks are converted to weights
wi = ri=n, where ri is the IDF rank of the ith token. The least frequent (highest
IDF) token has rank n, the most frequent token, rank 1. For example, in the
phrase \pupillary miosis" of the right eye, the weighted word vector would be:
[pupillary:0.83, miosis:1.0, of:0.33, the:0.17, right:0.5, eye:0.67]. The weights are
used in calculating weighted cosine similarity between the mention tokens and
each candidate CUI variant. All candidates (up to 100) having variants with
similarity to the mention text above a threshold are kept for the reranking step.
Converting IDF values to rank weights normalizes and smoothes the IDF values.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Candidate Reranking</title>
        <p>The candidate CUIs are reranked by measuring the similarity between mention
context and candidate context. The mention context is a relatively large window
of text (averaging 6-7 sentences) surrounding the mention. Both the mention
context and the candidate context are treated as bags of words in computing
the cosine similarity. We considered di erent context window sizes. The results
reported below use the full sentence containing the mention span (Sm). On the
candidate CUI side, we implemented following three context generators:
Gloss-Based Medical Concept Resolution (gbmcr) Two contributing sources
in UMLS are MeSH and NCI, which together contribute de nitions for only
roughly 3% of the concepts. Nevertheless, for 86% of the medical concept
mentions in our experiments, at least one of the ltered candidate concepts had at
least one MeSH or NCI de nition. In gbmcr, candidates are ranked according to
the cosine similarity between the words in the mention span (Sm) and the words
from the MeSH de nition of the candidate, if one exists, or the words from the
NCI de nition.</p>
        <p>Neighbor-Based Medical Concept Resolution (nbmcr) In addition to
taxonomic relations, UMLS contains semantic relations. In nbmcr, we consider
\neighbor concepts" related to the candidate concept via a subset of \clinically
relevant" relations based on the semantic type of the candidate CUI. For
example, for candidates of type Disease or Syndrome, we consider related symptoms,
treatments, risk factors, etc. Candidates are ranked according to the similarity
between the words in Sm and the words in variants of the neighbor CUIs
semantically related to the candidate.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Variants-Based Medical Concept Resolution The bag of words of all of</title>
        <p>the variant labels in UMLS of the candidate CUI make up the vbmcr candidate
context.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <p>In order to evaluate our MCR methods, we compare their performance to MetaMap
(mmap)2 on a dataset that contains 1,570 medical term spans extracted from 100
short descriptions (averaging roughly 8 sentences, 100 words) of patient
scenarios. The MCR algorithms can produce a ranked list of as many CUIs as there are
ltered candidates (arbitrarily capped at 100). We included the top three ranked
concepts for each factor in the evaluation. Five human judges were randomly
assigned roughly 560 factor-CUI mappings each. The assignments did not overlap,
but judges also rated 41 mappings in common. The average pairwise kappa for
judge agreement was 0.6. For each CUI found for a medical term, judges gave
a score of: 0 (inappropriate for the factor), 1 (appropriate for the factor) and
2 (appropriate for the factor and better than concepts scoring 1 for the same
factor). The purpose of distinguishing two grades of appropriate is to verify one
of the original motivations for MCR: that even when a CUI appropriate to the
exact span exists, there are often more speci c CUIs that are more appropriate
2 MetaMap: http://metamap.nlm.nih.gov/
considering more context. For example, consider the sentence \She has pain in
the epigastric region and sometimes on the right side of her abdomen", here, the
CUI for \pain" is appropriate, but UMLS also has CUIs for \abdominal pain"
and \right sided abdominal pain".</p>
      <p>mmap
gbmcr
nbmcr
vbmcr</p>
      <p>Best-is-correct Appropriate-is-correct
Precision Recall F1 Precision Recall F1
0.614 0.608 0.611 0.794 0.787 0.790
0.450 0.427 0.438 0.628 0.596 0.611
0.567 0.538 0.552 0.624 0.592 0.608
0.721 0.685 0.703 0.762 0.723 0.742</p>
      <p>Table 1. Performance
3.1</p>
      <sec id="sec-3-1">
        <title>Results and Discussion</title>
        <p>We calculate the precision, recall and F1-measure by considering two settings:
best (CUIs scoring 2 are correct; CUIs scoring 1 are correct only if there are
no CUIs scoring 2); and appropriate (CUIs scoring 2 or 1 are correct). We can
compare the performance of di erent methods in a strict environment and a
more relaxed one. Table 1 shows that our method vbmcr outperforms all other
approaches in strict environment (Best-is-correct). Particularly, vbmcr achieved
more than 13% improvement over state of the art method of medical entity
linking i.e. MetaMap. The other two MCR methods (nbmcr and gbmcr) did
not perform as well. MetaMap achieved the highest scores in the relaxed setting
(Appropriate-is-correct). This shows that MCR is able to nd more speci c
concepts and takes into consideration more context; MetaMap performs well
when there is a close match between the mention text and CUI variants, and no
more speci c CUIs exist.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4 Conclusion</title>
      <p>We introduced the notion of Medical Concept Resolution (MCR), a knowledge
base lookup task in which terms expressed in medical text are identi ed in
a knowledge base. We argued that MCR is more di cult than standard entity
linking problems because medical terms themselves are far more composable and
contextual, and determining the correct span of text to search for in a knowledge
base is more complex. We introduced three aspects of this complexity: discovery,
multiplicity, and granularity. Further, we presented a set of new algorithms for
performing MCR and showed that our methods outperformed state-of-the-art
methods.</p>
      <p>References</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>N.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Buitelaar</surname>
          </string-name>
          .
          <article-title>Wikipedia-based distributional semantics for entity relatedness</article-title>
          .
          <source>In 2014 AAAI Fall Symposium Series</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>H.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Grishman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H. T.</given-names>
            <surname>Dang</surname>
          </string-name>
          .
          <article-title>Overview of the tac 2010 knowledge base population track</article-title>
          .
          <source>In TAC</source>
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Fan</surname>
          </string-name>
          .
          <article-title>Medical relation extraction with manifold models</article-title>
          .
          <source>In ACL</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>