<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>March</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Interactive Topic Model with Enhanced Interpretability</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jun Wang</string-name>
          <email>jun.wang@us.fujitsu.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Junfu Xiang</string-name>
          <email>xiangjf.fnst@cn.fujitsu.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Changsheng Zhao</string-name>
          <email>cz2458@columbia.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kanji Uchino</string-name>
          <email>kanji@us.fujitsu.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Columbia University</institution>
          ,
          <addr-line>New York City, NY</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Fujitsu Laboratories of America</institution>
          ,
          <addr-line>Sunnyvale, CA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Fujitsu Nanda Software Tech. Co., Ltd.</institution>
          ,
          <addr-line>Nanjing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>20</volume>
      <issue>2019</issue>
      <abstract>
        <p>Although existing interactive topic models allow untrained end users to easily encode their feedback and iteratively reifne the topic models, their unigram representations often result in ambiguous description of topics and poor interpretability for users. To address the problems, this paper proposes the first phrase-based interactive topic model which can provide both high interpretability and high interactivity with human in the loop. First, we present an approach to augment unigrams with a list of probable phrases which ofers a more intuitively interpretable and accurate topic description, and further eficiently encode users' feedback with phrase constraints in interactive processes of refining topic models. Second, the proposed approach is demonstrated and examined with real data.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
      <p>• Human-centered computing → Human computer
interaction (HCI); • Computing methodologies →
Machine learning.</p>
      <p>IUI Workshops’19, March 20, 2019, Los Angeles, USA
© Copyright 2019 for the individual papers by the papers’ authors. Copying
permitted for private and academic purposes. This volume is published and
copyrighted by its editors.</p>
    </sec>
    <sec id="sec-2">
      <title>1 INTRODUCTION</title>
      <p>
        Topic models are a useful and ubiquitous tool for
understanding large electronic archives, which can be used to discover
the hidden themes that pervade the collection and annotate
the documents according to those themes, and further
organize, summarize, and search the texts [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. However, as
fully-unsupervised methods, vanilla topic models, such as
Latent Dirichlet allocation (LDA) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], often generate some
topics which do not fully make sense to end users [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Some
generated topics may not well correspond to meaningful
concepts, for instances, two or more themes can be confused
into one topic or two diferent topics can be (near) duplicates.
Some topics may not align well with user modeling goals
or judgements. For many users in computational social
science, digital humanities, and information studies, who are
not machine learning experts, topic models are often a “take
it or leave it” proposition [
        <xref ref-type="bibr" rid="ref10 ref6">6, 10</xref>
        ]. Diferent from purely
unsupervised topic models that often result in unexpected topics,
taking prior knowledge into account enables us to produce
more meaningful topics [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Interactive topic models with
human in the loop are proposed and allow untrained end
users to easily encode their feedback as prior knowledge
and iteratively refine the topic models (e.g., changing which
words are included in a topic, or merging or splitting topics)
[
        <xref ref-type="bibr" rid="ref10 ref12 ref16">10, 12, 16</xref>
        ].
      </p>
      <p>
        A topic is typically modeled as a categorical distribution
over terms, and frequent terms related by a common theme
are expected to have a large probability [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. It is of interest
to visualize these topics in order to facilitate human
interpretation and exploration of the large amounts of unorganized
text, and a list of most probable terms is often used to
describe individual topics. Similar to vanilla topic models, all
existing interactive topic models are represented with
unigrams, which often provide ambiguous representation of
the topic and poor interpretability for end users [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Smith
et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] conducted user studies on a unigram-based
interactive topic model, and also were aware of the requests from
participants for the ability to add phrases and support of
multi-word refinements as opposed to single tokens.
      </p>
      <p>
        As shown in Table 1, human interpretation often relies
on inherent grouping of words into phrases, and
augmenting unigrams with a list of probable phrases ofers a more
intuitively interpretable and accurate topic description [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
Under the ‘bag-of-words’ assumption of unigrams, phrases
are decomposed and a phrase’s meaning may be lost, so topic
models need to systematically assign topics to whole phrases.
Several phrase-based topic models [
        <xref ref-type="bibr" rid="ref19 ref3 ref7 ref8">3, 7, 8, 19</xref>
        ] have been
proposed to discover topical phrases and address the prevalent
deficiency in visualizing topics using unigrams. But all these
models are static systems which end users cannot easily and
interactively refine, so they have the same “take it or leave
it” issues.
      </p>
      <p>To address the above problems, this paper proposes the
ifrst phrase-based interactive topic model which can provide
both high interpretability and high interactivity as shown in
Figure 1. First, we present an approach to discover topical
phrases with mixed lengths by detecting phrases and
phrasebased topic inference, and further eficiently encode users’
feedback with phrase constraints into interactive processes
of refining topic models. Second, the proposed approach is
demonstrated and examined with real data.</p>
      <p>We organize the remainder of the paper as follows.
Section 2 introduces some related work. Section 3 illustrates
the general framework we propose. Section 4 presents our
experimental results on real-world data. Finally, section 5
summarizes our work and discuss the future work.
2</p>
    </sec>
    <sec id="sec-3">
      <title>RELATED WORK</title>
      <p>
        Various approaches have been proposed to encode users’
feedback as prior knowledge into topic models instead of
purely relying on how often words co-occur in diferent
contexts. Andrzejewski et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] imposed Dirichlet Forest prior
over the topic-word categoricals to encode the Must-Links
and Cannot-Links between words. Words with Must-Links
are encouraged to have similar probabilities within all topics
while those with Cannot-Links are disallowed to
simultaneously have large probabilities within any topic. Xie et al.
[
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] studied how to incorporate the external word
correlation knowledge to improve the coherence of topic modeling,
and built a Markov Random Field (MRF) regularized topic
model encouraging words labeled as similar to share the
same topic label. Yang et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] integrated lexical
association into topic optimization using tree priors to improve
topic interpretability, which provided a flexible framework
that can take advantage of both first order word
associations and the higher-order associations captured by word
embeddings.
      </p>
      <p>
        Several unigram-based interactive topic models have been
proposed and studied. Hu et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] extended the framework
of Dirichlet Forest prior and proposed the first interactive
topic model. Lee et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] employed a user-centered
approach to identify a set of topic refinement operations that
users expect to have in a interactive topic model system.
However, they did not implement underlying algorithm to
refine topic models and only used Wizard-of-Oz refinements:
the resulting topics were updated superficially—not as the
output of a data-driven statistical model (the goal of topic
models) [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Smith et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] further implemented an
eficient asymmetric prior-based interactive topic model with
a broader set of user-centered refinement operations, and
conducted a study with twelve non-expert participants to
examine how end users are afected by issues that arise with
a fully interactive, user-centered system.
      </p>
      <p>
        Some researchers proposed various phrase-based topic
models. Wang et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] attempted to infer phrases and
topics simultaneously by creating complex generative
mechanism. The resultant models can directly output phrases
and their latent topic assignment. It used additional latent
variables and word-specific multinomials to model bi-grams,
and these bigrams can be combined to form n-gram phrases.
KERT [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and Turbo Topics [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] constructed topical phrases as
a post-processing step to unigram-based topic models. These
methods generally produce low-quality topical phrases or
sufer from poor scalability outside small datasets [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
ElKishky et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and Wang et al. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] proposed a
computationally eficient and efective approaches, which combines
a phrase mining framework to segment a document into
single and multi-word phrases, and a topic model with phrase
constraints that operates on the induced document partition.
3
      </p>
    </sec>
    <sec id="sec-4">
      <title>FRAMEWORK</title>
      <p>
        For phrase-based topic models, the better method is first
mining phrases and segmenting each document into single and
multiword phrases, and then running topic inference with
phrase constraints [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. End users can give feedback using a
high
It
n
e
r
a
c
t
ii
v
t
y
Low
variety of refinement operations on topical phrase
visualization, and users’ feedback will update the prior knowledge
and the phrase-based topic inference will be rerun based on
the updated prior. As shown in Figure 2, the process forms
a loop in which users can continuously and interactively
update and refine the topic model with phrase constraints.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Phrase Mining</title>
      <p>
        Phrase mining is a text mining technique that discovers
semantically meaningful phrases from massive text. Recent
data-driven approaches opt instead to make use of frequency
statistics in the corpus to address both candidate generation
and quality estimation [
        <xref ref-type="bibr" rid="ref13 ref15 ref18 ref7">7, 13, 15, 18</xref>
        ]. They do not rely on
complex linguistic feature generation, domain-specific rules
or extensive labeling eforts. Instead, they rely on large
corpora containing hundreds of thousands of documents to help
deliver superior performance several indicators, including
frequency, mutual information, branching entropy and
comparison to super/sub-sequences, were proposed to extract
n-grams that reliably indicate frequent, concise concepts
[
        <xref ref-type="bibr" rid="ref13 ref15 ref18 ref7">7, 13, 15, 18</xref>
        ].
      </p>
    </sec>
    <sec id="sec-6">
      <title>Phrase-based topic inference</title>
      <p>
        After inducing a partition on each document, we perform
topic inference to associate the same topic to each word in a
phrase and thus naturally to the phrase as a whole. El-Kishky
et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] proposed a probabilistic graphical model PhraseLDA
based on a generative process almost same as LDA but with
constraints on topics of phrases, and corresponding
phrasebased topic inference can be smoothly updated from
unigrambased topic inference of LDA.
      </p>
      <p>LDA assumes that a document may contain multiple topics,
where a topic is defined to be a categorical distribution over
words in the vocabulary. The generative process is as follows:
(1) Draw ϕk ∼ Dirichlet (β), for 1 ≤ k ≤ K
(2) For document d, where 1 ≤ d ≤ D:
(a) Draw θk ∼ Dirichlet (α )
(b) For n-th word in document d, where 1 ≤ n ≤ Nd
(i) Draw zd,n ∼ Cateдorical (θd )
(ii) Draw wd,n ∼ Cateдorical (ϕzd,n )
α is a K -dimensional vector (α1, . . . , αK ), and β is a V
dimensional vector (β1, . . . , βV ). K is the number of topics,
D is the number of documents, V is the size of vocabulary,
and Nd is the number of words in the document d.</p>
      <p>Based on its generative process, the joint distribution of
LDA (1) can be represented as the product of two
DirichletMultinomial distributions (2). The Dirichlet-Multinomial
expressions (3) can be further simplified using the feature of
gamma function (represented by Γ) later.</p>
      <p>p(W , Z ; α, β)
∫
p(W , Z, Θ, Φ; α, β)dΘdΦ</p>
      <p>∫
p(Z, Θ; α )dΘ ×</p>
      <p>p(W , Φ|Z ; β)dΦ
=
=
∝
∫
×
= p(Z ; α ) × p(W |Z ; β)
= Dir Mult (Z ; α ) × Dir Mult (W |Z ; β)</p>
      <p>D
Ö
ÎK</p>
      <p>k=1 Γ(Nd,k + αk )
d=1 Γ(ÍkK=1(Nd,k + αk ))</p>
      <p>K
Ö
ÎV</p>
      <p>v=1 Γ(Nk,v + βv )
k=1 Γ(ÍvV=1(Nk,v + βv ))
(1)
(2)
(3)</p>
      <p>W is the collection of all words in D documents, and Z is
the collection of corresponding topics assigned to each word
in W . Θ is the collection of (θ1, . . . , θK ), and Φ is the
collection of (ϕ1, . . . , ϕK ). Nd,k is the number of words assigned to
topic k in the document d, and Nk,v is the number of words
with topic k and value v in the vocabulary.</p>
      <p>Because the generative process of PhraseLDA is almost
same as LDA except phrase constraints on topics, its joint
distribution is same as LDA in the above (3). But PhraseLDA
and LDA are diferent in calculating the full conditional
distribution (4), by which we can sample topics using Gibbs
sampling. And we know that the full conditional distribution
(4) is proportional to the joint distribution (1).</p>
      <p>p(za,b = i |Z¬a,b, W ; α, β)
= p(za,b = i |wa,b = j, Z¬a,b, W¬a,b ; α, β)
∝ p(W , Z ; α, β)
za,b is the topic assigned to the wa,b , which is the b-th
unit in the document a. W¬a,b is the collection of all units
except wa,b , and Z¬a,b is the collection of corresponding
topics assignments. In LDA wa,b is the b-th word in the
document a, and in PhraseLDA wa,b is the b-th phrase in
the document a, and this diference results in diferent topic
inference processes.</p>
      <p>For PhraseLDA, we can simplify two Dirichlet-Multinomial
expressions (3) to sample the topic of a phrase as follows,
and please see Appendix for detail derivations.
Öla,b (Na¬,ai ,b + αi + д − 1) × (Ni¬,wa,ab,b,д + βwa,b,д )
д − 1 + ÍV</p>
      <p>v=1(Ni¬,va,b + βv )
la,b is the length of the b-th phrase in the document a, and
wa,b,д is the д-th word in phrase wa,b . Na¬,ai ,b is the number
of words assigned to topic i in the document a after excluding
the phrase wa,b , and Ni¬,va,b is the number of words with topic
i and value v after excluding the phrase wa,b .</p>
      <p>
        α and β can be optimized using the method presented by
Minka [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] for the phrase-based topic model before
refinement operations of users.
      </p>
    </sec>
    <sec id="sec-7">
      <title>Refinement Operations of Users</title>
      <p>
        Smith et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] identified a set of refinements that users
expected to be able to use in a interactive topic model, and
implemented seven refinements requested by users: add word,
remove word, change word order, remove document,
split topic, merge topic, and add to stop words.
      </p>
      <p>
        Participants of the qualitative evaluation in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] found
change word order to be one of the least useful refinements,
and as shown in Table 1, with the phrase representation of
topics the phrase order does not have much influence on
human interpretability. Add to stop words is easy, and we
just exclude the word w from the vocabulary and ensures that
the Gibbs sampler ignores all occurrences of w in the corpus.
So we can skip detail discussions of these two operations
in the paper, and extend other operations based on phrases
instead of words.
      </p>
    </sec>
    <sec id="sec-8">
      <title>Update Prior Knowledge with phrase constraints</title>
      <p>
        Adding a human in the loop requires the user to be able
to inject their knowledge via feedback into the sampling
equation to guide the algorithm to better topics [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>(4)
(5)</p>
      <p>
        Dirichlet Forest prior has been widely used to encode
users’ feedback as prior knowledge in various interactive
topic models [
        <xref ref-type="bibr" rid="ref1 ref10 ref16">1, 10, 16</xref>
        ]. This kind of priors attempted to
enforce hard and topic-independent rules that similar words
should have similar probabilities in all topics, which is
questionable in that two words with similar representativeness
of one topic are not necessarily of equal importance for
another topic [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. For example, in the fruit topic, the words
apple and orange have similar representativeness, while in
an IT company topic, apple has much higher importance
than orange. Dirichlet Forest prior is unable to diferentiate
the subtleties of word sense across topics and would falsely
put irrelevant words into the same topic [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. For instance,
since orange and Microsoft are both labeled as similar to
apple and are required to have similar probabilities in all
topics as apple has, in the end, they will be unreasonably
allocated to the same topic.
      </p>
      <p>
        Wallach et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] has found that an asymmetric Dirichlet
prior has substantial advantages over a symmetric prior in
topic models, and to address the above problems, Smith et
al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] proposed an asymmetric prior which encodes users’
feedback through modifying the Dirichlet prior parameters
for each document and each topic involved. Similar idea can
be extended to address phrase constraints and applied to
phrase-based interactive model. In the previous section on
phrase-based topic inference, all documents share the same
α and all topics share the same β. Here, every document a
and every topic i involved in refinement operations need
corresponding separate α (a) and β(i), respectively, and the
sampling equation (5) should be updated as follows:
Öla,b (Na¬,ai ,b + αi(a) + д − 1) × (Ni¬,wa,ab,b,д + β w(ia),b,д )
д=1
д − 1 + ÍV i
v=1(Ni¬,va,b + βv( ))
      </p>
      <p>(6)</p>
      <p>
        These priors α (a) and β(i) are sometimes called
“pseudocounts” [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], and interactive models can take advantage of
them by creating pseudo-counts to encourage the changes
users want to see in the topic [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>
        Remove document and Merge topic are
straightforward and almost same as the unigram-based updates
proposed in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>• Remove document: to remove the document a from
topic i, we invalidate the topic assignment for all words
in the document a and assign a very small prior αi(a)
to the topic i in a.
• Merge topic: merging topics i1 and i2 means the model
will have a combined topic that represents both i1 and
i2. We assign i1 to all words that were previously
assigned to i2, and reduce the number of topics.</p>
      <p>
        For Remove phrase, Split topic and Add phrase,
corresponding updates are a bit more complicated since we
need to deal with phrase constraints. For a phrase p, lp is the more than 50 times of our system. We also tried to extend
length of p and pд is д-th word in p where 1 ≤ д ≤ lp . the model presented in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] to support phrases, and check
• Remove phrase: to remove the phrase p from topic if human interpretability of generated topic are improved.
i, we need to locate all occurrences of p assigned to Correlation scores based on phrase embedding vectors
gentopic i and invalidate their topic assignment. For topic erated by Fasttext [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] are calculated to build two-level tree
prior. The model attempts to encourage phrases close in
emi, very small prior βp(iд) is assigned to each word pд bedding vector space to appear in the same topic, but we
contained in p. found that it only performs slightly better on downstream
• Split topic: to split topic i1 the user provides a subset tasks, such as classification, and does not really help to
enof seed phrases, which need to be moved from the orig- hance human interpretability. The above observations led to
inal topic i, to a new topic i2. We invalidate the original creating our current system.
topic assignment of all seed phrase occurrences, in- Our phrase-based topic model before refinement
operacrease the number of topics, and assign large prior βp(iд2) tions was initialized with 2000 iterations using the optimized
for each word pд contained in each seed phrase p for α (with mean 0.415) and β (with mean 0.015). Since this is
the new topic i2. a one-time job, we can set an even larger iteration number.
• Add phrase: to add the phrase p to topic i, we inval- The number of sampling iterations for updating and refining
idate all occurrences of p from all other topics and model can be tuned according to latency acceptable for users
encourage the Gibbs sampler to assign topic i for each (for example, less than 1 minute), and we set the number
occurrence, and we increase the prior of each word as 400. Similar to [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], βp(iд) is set as 0.000001 for remove
contained in p for topic i. phrase and split topic.
      </p>
      <p>Since this paper focuses on improving human
interpretabil4 EXPERIMENTS ity of interactive topic models, and as we have known, the
We deployed the phrase-based interactive topic models as automated methods of measuring topic quality in terms of
a part of our corporate learning platform for data scientist coherence often do not correlate well with human
judgetraining programs, in which a database contains 19852 recent ment and interpretation, and in addition, these methods are
machine learning related papers collected from ICML/NIPS/arXiv. generally only available for unigram-based models, so the</p>
      <p>
        For phrase mining, we used our own tool based on gen- experimental evaluation in this paper are mainly based on
eralized sufix tree (GST) presented in [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], and segmented user studies. 5 participants with computer science or
electhe titles and abstracts of all papers into a collections of tronic engineering background, who are users of the
cormeaningful phrases. porate learning platform, were asked to use and refine the
      </p>
      <p>In order to facilitate human exploration and interpretation, phrase-based interactive topic model.
we visualize these papers into 20 topics using our system, Our user studies showed the split topic and remove
and learners can further interactively refine the topics using phrase are the most commonly used operations, and
octheir domain knowledge as shown in Figure 3. A list of topics casionally merge topic is used based on users’ personal
in the left panel are represented by top three phrases of each preference. But add phrase is a relatively rare operation,
topic. Selecting a topic displays more detail in the right panel: because in most cases it is not easy for users to discover or
the top 30 phrases with frequency and top associated docu- remember phrases not presented to them, especially for a
ments with corresponding percentage. Users can click and new domain.
select phrases for removing with remove phrase button or There are a couple of coherent but non-informative
topfor splitting with split topic button, click and select docu- ics. For example, one topic mainly contains phrases such as
ments for removing with remove document button, add training data, data sets, data points, and another topic mainly
new phrases from the vocabulary with add phrase button, contains phrases such as experimental results, theoretical
reselect phrases and click the add to stop words button to sults. Except these uninformative topics, all 5 participants
move to the stop words list, or click merge topic button to agreed that our system can significantly refine quality and
coinput two topics for merging. herence of all other topics and consistently improve human</p>
      <p>
        Before we implemented our phrase-based interactive model interpretability of topic modeling. The user studies showed
illustrated in Figure 2, we first tried the model based on that a well-organized structure can be established and refined
Dirichlet Forest prior presented in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and found a few by our phrase-based interactive topic model.
drawbacks. Instead of direct modification, people are forced Several typical examples from participants’ real
refineto think of pairwise relation, which is counter-intuitive. Its ment operations are demonstrated here. In the first example
prior tree structure is hard to encode phrase constraints and shown in Figure 4, a participant found that two unrelated
results in an extremely slow convergence, whose latency is
topics (social media and Autonomous driving) were
mistakenly mixed into one topic, and she selected Social media as
a seed phrase for split topic. In the second example shown
in Figure 5, the existing topic was actually fine, but a
participant wanted to refine and separate a fine-grained new
topic on face recognition from the existing topic on image
processing, and she selected Face recognition as a seed phrase
for split topic. Interestingly, although only one seed phrase
was selected for the new topic in the above two examples,
other unselected phrases related to the seed can correctly
move to the new topic as well. In the third example, a
participant found that an important phrase Computer vision was
assigned to a unexpected and inappropriate topic which is
not really meaningful, and she wanted to remove Computer
vision from this inappropriate topic and check if it is possible
to finally move it to a meaningful topic. After two rounds
of remove phrase, the phrase Computer vision moved to an
appropriate topic as shown in Figure 6.
      </p>
    </sec>
    <sec id="sec-9">
      <title>5 CONCLUSION AND FUTURE WORK</title>
      <p>This paper proposes the first phrase-based interactive topic
model which can provide both high interpretability and high
interactivity with human in the loop, and demonstrates and
examines the proposed approach with real data. Although
Select a seed phrase
“social media” for split
topic operation
Social media
Recommender systems
Autonomous driving
Autonomous vehicles
Traffic sign
Fake news
User study
User preferences
Differential privacy
Shed light
Social sciences
Differentially private
Traffic light</p>
      <p>
        Social media
Recommender systems
User study
Past decade
Fake news
User preferences
Differential privacy
Social sciences
Differentially private
Research topic
Autonomous driving
Autonomous vehicles
Traffic sign
Improve generalization
Specifically designed
Aerial vehicles
Autonomous cars
Broad class
Fully automatic
Designed specifically
the latency of our system has significantly improved
compared with previous systems based on tree prior, it can still be
a major issue for large scale data, so we need to study more
eficient methods of inference using sparsity [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], which can
be smoothly applied to systems with phrase constraints.
Current methods for automatically measuring topic coherence
and quality are also mainly for models based on unigrams
Select a seed phrase
“face recognition” for
split topic operation
Image classification
Input image
Face recognition
Single image
Style transfer
Image captioning
Face images
Image processing
Image retrieval
Natural images
Facial expressions
Image quality
Compressed sensing
      </p>
      <p>
        Face recognition
Style transfer
Face images
Facial expressions
Facial landmark
Facial attributes
Pattern recognition
Face detection
Face verification
Referring expression
Image classification
Input image
Single image
Medical imaging
Image captioning
Image processing
Image generation
Generated images
Natural images
MR images
[
        <xref ref-type="bibr" rid="ref11 ref2">2, 11</xref>
        ], so we also need to study how to extend corresponding
methods for phrase-based models.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>David</given-names>
            <surname>Andrzejewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Xiaojin</given-names>
            <surname>Zhu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Mark</given-names>
            <surname>Craven</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Incorporating Domain Knowledge into Topic Modeling via Dirichlet Forest Priors</article-title>
          .
          <source>In Proceedings of the 26th Annual International Conference on Machine Learning (ICML '09)</source>
          .
          <fpage>25</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Shraey</given-names>
            <surname>Bhatia</surname>
          </string-name>
          , Jey Han Lau, and
          <string-name>
            <given-names>Timothy</given-names>
            <surname>Baldwin</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Topic Intrusion for Automatic Topic Model Evaluation</article-title>
          .
          <source>In EMNLP. Association for Computational Linguistics</source>
          ,
          <fpage>844</fpage>
          -
          <lpage>849</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Blei</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Laferty</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Visualizing Topics with Multi-Word Expressions</article-title>
          .
          <source>arXiv 0907.1013v1</source>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>David</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>Andrew Y.</given-names>
          </string-name>
          <string-name>
            <surname>Ng</surname>
            , and
            <given-names>Michael I.</given-names>
          </string-name>
          <string-name>
            <surname>Jordan</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>Latent Dirichlet Allocation</article-title>
          .
          <source>J. Mach. Learn. Res. 3 (March</source>
          <year>2003</year>
          ),
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Piotr</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          , Edouard Grave, Armand Joulin, and
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Enriching Word Vectors with Subword Information</article-title>
          .
          <source>arXiv preprint arXiv:1607.04606</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Jordan</given-names>
            <surname>Boyd-Graber</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Humans and Computers Working Together to Measure Machine Learning Interpretability</article-title>
          .
          <source>The Bridge</source>
          <volume>47</volume>
          (
          <year>2017</year>
          ),
          <fpage>6</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Marina</given-names>
            <surname>Danilevsky</surname>
          </string-name>
          , Chi Wang, Nihit Desai, Xiang Ren, Jingyi Guo, and Jiawei Han.
          <year>2014</year>
          .
          <article-title>Automatic Construction and Ranking of Topical Keyphrases on Collections of Short Documents</article-title>
          .
          <source>In Proceedings of the 2014 SIAM International Conference on Data Mining</source>
          , Philadelphia, Pennsylvania, USA, April
          <volume>24</volume>
          -
          <issue>26</issue>
          ,
          <year>2014</year>
          .
          <fpage>398</fpage>
          -
          <lpage>406</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Ahmed</given-names>
            <surname>El-Kishky</surname>
          </string-name>
          , Yanglei Song, Chi Wang,
          <string-name>
            <surname>Clare R. Voss</surname>
          </string-name>
          , and Jiawei Han.
          <year>2014</year>
          .
          <article-title>Scalable Topical Phrase Mining from Text Corpora</article-title>
          .
          <source>Proc. VLDB Endow</source>
          .
          <volume>8</volume>
          ,
          <issue>3</issue>
          (Nov.
          <year>2014</year>
          ),
          <fpage>305</fpage>
          -
          <lpage>316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Gregor</given-names>
            <surname>Heinrich</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Parameter estimation for text analysis</article-title>
          .
          <source>Technical Report.</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Yuening</surname>
            <given-names>Hu</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            Boyd-Graber,
            <given-names>Brianna</given-names>
          </string-name>
          <string-name>
            <surname>Satinof</surname>
            , and
            <given-names>Alison</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Interactive Topic Modeling</article-title>
          .
          <source>Machine Learning</source>
          <volume>95</volume>
          (
          <year>2014</year>
          ),
          <fpage>423</fpage>
          -
          <lpage>469</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Jey</given-names>
            <surname>Han</surname>
          </string-name>
          <string-name>
            <surname>Lau</surname>
          </string-name>
          , David Newman,
          <string-name>
            <given-names>and Timothy</given-names>
            <surname>Baldwin</surname>
          </string-name>
          .
          <year>2014</year>
          . Machine Reading Tea Leaves:
          <article-title>Automatically Evaluating Topic Coherence and Topic Model Quality</article-title>
          .
          <source>In EACL. The Association for Computer Linguistics</source>
          ,
          <fpage>530</fpage>
          -
          <lpage>539</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Tak</surname>
            <given-names>Y</given-names>
          </string-name>
          “eMoendiLcaelei,mAaglies”ornelaStmedith,
          <string-name>
            <surname>Kevin</surname>
            <given-names>Seppi</given-names>
          </string-name>
          , Niklas Elmqvist,
          <string-name>
            <surname>Jordan</surname>
            <given-names>BoydGrabteorp</given-names>
          </string-name>
          , iacnd Leah“FCionmdpluattaetrio.n2a0l1c7o.mTphleexitHy”uman Touch:
          <article-title>How Non-expert Users Perceive, I nretleartepdretotp,iacn</article-title>
          ,bdutFsitxillTuonpdeicsirMedodels.
          <source>International Journal of Human-Comp ufotre“rcSomtupduiteesr v(2is0io1n7”).</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Jialu</surname>
            <given-names>Liu</given-names>
          </string-name>
          , Jingbo Shang, Chi Wang,
          <string-name>
            <surname>Xiang Ren</surname>
          </string-name>
          , and Jiawei Han.
          <year>2015</year>
          .
          <article-title>Mining Quality Phrases from Massive Text Corpora</article-title>
          .
          <source>In Proceedings of the 2015 ACM SIGMOD International Conference on Management of Dat1ast(SroIuGnMdOD '15)</source>
          .
          <fpage>1729</fpage>
          -
          <lpage>1744</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>ThommoavisngP. Minka</surname>
          </string-name>
          .
          <year>20M0o0v</year>
          .
          <article-title>eEfrsotmimthaetipnrgevaiouDsirichlet distribution</article-title>
          .
          <source>Technical Report</source>
          . MIT. topic to a new topic
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Shang</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. R.</given-names>
            <surname>Voss</surname>
          </string-name>
          , and J. Han.
          <year>2018</year>
          .
          <article-title>Automated Phr2ansderoMunidning from Massive Text Corpora</article-title>
          .
          <source>IEEE Transactions on moving Knowledge and Data Engineering</source>
          <volume>30</volume>
          ,
          <issue>10</issue>
          (
          <year>2018</year>
          ),
          <fpage>1825</fpage>
          -
          <lpage>1837</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Alison</surname>
            <given-names>Smith,</given-names>
          </string-name>
          <article-title>VarUunndeKsuiremd ator</article-title>
          ,piJcofrodran Boyd-Graber,
          <string-name>
            <given-names>Kevin</given-names>
            <surname>Seppi</surname>
          </string-name>
          , and Leah Findlater.
          <volume>20</volume>
          “1c8o.
          <article-title>mpCultoersivnisgiont”he Loop: User-Centered Design and Evaluation of a Human-in-the-Loop Topic Modeling System</article-title>
          .
          <source>In 23rd International Conference on Intelligent User Interfaces (IUI '18)</source>
          .
          <fpage>293</fpage>
          -
          <lpage>304</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Hanna</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Wallach</surname>
          </string-name>
          , David Mimno, and
          <string-name>
            <surname>Andrew McCallum</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <string-name>
            <surname>Rethinking</surname>
            <given-names>LDA</given-names>
          </string-name>
          :
          <article-title>Why Priors Matter</article-title>
          .
          <source>In Proceedings of the 22Nd International Conference on Neural Information Processing Systems (NIPS'09)</source>
          .
          <fpage>1973</fpage>
          -
          <lpage>1981</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Jun</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Junfu Xiang</surname>
            , and
            <given-names>Kanji</given-names>
          </string-name>
          <string-name>
            <surname>Uchino</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Topic-Specific Recommendation for Open Education Resources</article-title>
          .
          <source>In Advances in Web-Based Learning - ICWL</source>
          <year>2015</year>
          ,
          <string-name>
            <surname>Frederick</surname>
            <given-names>W.B.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Ralf</given-names>
          </string-name>
          <string-name>
            <surname>Klamma</surname>
          </string-name>
          , Mart Laanpere, Jun Zhang, Baltasar Fernández Manjón, and
          <string-name>
            <surname>Rynson W.H. Lau</surname>
          </string-name>
          (Eds.). Springer International Publishing, Cham,
          <fpage>71</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Xuerui</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andrew McCallum</surname>
            ,
            <given-names>and Xing</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Topical NGrams: Phrase and Topic Discovery, with an Application to Information Retrieval</article-title>
          .
          <source>In Proceedings of the 2007 Seventh IEEE International Conference on Data Mining (ICDM '07)</source>
          .
          <fpage>697</fpage>
          -
          <lpage>702</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Pengtao</surname>
            <given-names>Xie</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Diyi</given-names>
            <surname>Yang</surname>
          </string-name>
          , and
          <string-name>
            <surname>Eric P Xing</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Incorporating Word Correlation Knowledge into Topic Modeling. In The 2015 Conference of the North American Chapter of the Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Weiwei</surname>
            <given-names>Yang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            Boyd-Graber, and
            <given-names>Philip</given-names>
          </string-name>
          <string-name>
            <surname>Resnik</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Adapting Topic Models using Lexical Associations with Tree Priors</article-title>
          .
          <source>In Empirical Methods in Natural Language Processing.</source>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Yi</surname>
            <given-names>Yang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Doug</given-names>
            <surname>Downey</surname>
          </string-name>
          , and
          <string-name>
            <surname>Jordan</surname>
          </string-name>
          Boyd-Graber.
          <year>2015</year>
          .
          <article-title>Eficient Methods for Incorporating Knowledge into Topic Models</article-title>
          .
          <source>In Empirical Methods in Natural Language Processing.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>