<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Number of Topics Optimization: Clustering Approach*</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gazpromneft STC</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Moika River emb.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Saint Petersburg</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Russia krasnov.fv@gazprom-neft.ru</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Saint Petersburg State University</institution>
          ,
          <addr-line>7-9 Universitetskaya Emb., Saint Petersburg, 199034</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1949</year>
      </pub-date>
      <abstract>
        <p>Although topic models have been used to build clusters of documents for more than ten years, there is still a problem of choosing the optimal number of topics. The authors analyzed many fundamental studies undertaken on this subject in recent years. The main problem is the lack of a stable metric of the quality of topics obtained during the construction of the topic model. The authors analyzed the internal metrics of the topic model: Coherence, Contrast and Purity to determine the optimal number of topics and concluded that they are not applicable to solve this problem. The authors analyzed the approach to choosing the optimal number of topics based on the quality of the clusters. For this purpose, the authors considered the behavior of the cluster validation metrics: Davies Bouldin Index, Silhouette Coe cient, and CalinskiHarabaz. The cornerstone of the proposed new method of determining the optimal number of topics based on the following principles: { Setting up a topic model with additive regularization (ARTM) to separate noise topics; { Using dense vector representation (GloVe, FastText, Word2Vec); { Using a cosine measure for the distance in cluster metric that works better on vectors with large dimensions than Euclidean distance. The methodology developed by the authors for obtaining the optimal number of topics was tested on the collection of scienti c articles from the OnePetro library, selected by speci c themes. The experiment showed that the method proposed by the authors allows assessing the optimal number of topics for the topic model built on a small collection of Englishlanguage documents.</p>
      </abstract>
      <kwd-group>
        <kwd>clustering &amp; additive regularization topic model &amp; valida- tion metrics &amp; Davies Bouldin Index &amp; ARTM</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Topic models have been using successfully for clustering texts for many years.
One of the most common approaches to topic modeling is the Latent Dirichlet
Allocation (LDA) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] which models a xed number of topics selected as a
parameter based on the Dirichlet distribution for words and documents. The result
is a at, soft probabilistic clustering of terms by topics and documents by topics.
All the topics received are equal, they do not create any characteristic signs that
could help the researcher to identify the most useful topics, that is, to choose a
subset of topics that are best suited for human interpretation. The problem of
nding the metric characterizing such interpretability is a subject of study by
many researchers [
        <xref ref-type="bibr" rid="ref21 ref27 ref41 ref44">44, 27, 41, 21</xref>
        ].
      </p>
      <p>
        The topic model is not able to read the insights of the researcher and therefore
must have the settings for the task that the researcher is going to solve. According
to studies [
        <xref ref-type="bibr" rid="ref1 ref7">7, 1</xref>
        ] topic models based on the LDA have the following parameters:
: the parameter of the prior Dirichlet distribution for \documents-topics";
: parameter of the prior Dirichlet distribution for \topics-words";
tn: the number of topics;
b: the number of discarded initial iterations according to Gibbs sampling;
n: the number of samples;
si: sampling interval.
      </p>
      <p>
        In the recent study [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], published in 2018, an attempt was made to nd
the optimal values of the above parameters using the algorithm of Di erential
Evolution [
        <xref ref-type="bibr" rid="ref42">42</xref>
        ]. The authors chose a modi ed Jaccard Similarity metric as the
cost-function. As a result, a new LDADE algorithm was created, in which free
parameters from the Di erential Evolution algorithm appeared and they also
need to be optimized.
      </p>
      <p>
        There is a di erence between evaluating of a complete set of topics and
evaluating individual topics to lter out unwanted information (noise). To evaluate
a complete set of topics, researchers usually look at the Perplexity metric [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] for
the corpus of documents.
      </p>
      <p>
        This approach does not work very well according to the results of studies [
        <xref ref-type="bibr" rid="ref15 ref45">45,
15</xref>
        ] because the Perplexity does not have an absolute minimum, and with
increasing of iterations it becomes asymptotic [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
      </p>
      <p>The most common use of Perplexity is to detect the \elbow e ect", that is,
when the pattern of growth in the orderliness of the model changes drastically.
Perplexity depends on the power of the dictionary and the frequency distribution
of words in the collection, hence we get its drawbacks:
{ it cannot evaluate the quality of deletion of stop words and non-topic words;
{ it cannot compare rarefying methods for dictionary;
{ it cannot compare uni-gram and n-gram models.</p>
      <p>
        The authors of the LDA made a study of the quality of topics using the
Bayesian approach in [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ]. It important to note that the Hierarchical Dirichlet
process (HDP) [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ] solved the issue of the optimal number of topics, although it
used not for documents, but for the whole collection.
      </p>
      <p>
        Let us pay attention to the di erence between the LDA, HDP, and
hierarchical Latent Dirichlet Allocation (hLDA) [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ], since these are di erent topic
models. LDA creates a at, soft probabilistic clustering of terms by topic and
documents by topic. In the HDP model, instead of a xed number of topics for a
document, the Dirichlet process generates the number of topics, which leads to
the fact that the number of topics is also a random variable. The \hierarchical"
part of the name belongs to another level added by the Dirichlet process, which
creates several topics, and the topics themselves are still at clusters. The hLDA
model is an adaptation of the LDA, which models the topics as the distribution
of a new, predetermined number of topics taken from the Dirichlet distribution.
The hLDA model still considers the number of topics as a hyper parameter, that
is, regardless of the data. The di erence is that clustering is now hierarchical:
the hLDA model studies the clustering of the rst set of topics, providing more
general abstract relationships between topics (and, therefore, words and
documents). Note that all three models described (LDA, HDP, hLDA) add a new set
of parameters that require optimization, as is noted in the study [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        One of the main requirements for topic models is human interpretability [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ].
In other words, whether the topics contain words that, according to a person's
subjective judgments, are representative of a single coherent concept. In [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ],
Newman showed that the human assessment of interpretability well correlates
with an automated quality measure called coherence.
      </p>
      <p>
        The research [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] of 2018, proposed to minimize the Renyi and Tsallis
entropies to nd the optimal number of topics in the topic modeling. In this
study, topic models derived from large collections of texts are considered as
non-equilibrium complex systems, where the number of topics is considered as
the equivalent of temperature. This allows us to calculate the free energy of such
systems | the value through which the Renyi and Tsallis entropies are easily
expressed. The metrics obtained based on entropy make it possible to nd a
minimum depending on the number of topics for large collections, but in practice
we rarely nd small collections of documents.
      </p>
      <p>
        A study [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], published in 2018, proposed a matrix approach to improving the
accuracy of determining topics without using optimization. On the other hand,
the study [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] noted that increasing the accuracy of the model is contrary to
human interpretability. In particular, the study [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], completed in 2018, created
the VisArgue framework designed to visualize the model's learning process to
determine the most explainable topics.
      </p>
      <p>
        The use of the statistical measure TF{IDF as a metric for quantifying the
quality of topics was studied in [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ]. There is also a series of studies combining
the advantages of topic models and dense representations of word-vectors [
        <xref ref-type="bibr" rid="ref17 ref3 ref30 ref36">3, 30,
17, 36</xref>
        ].
      </p>
      <p>
        The motivation of the research conducted by the authors of this paper was
the fact that the study of a stable metric for the quality of topics continues.
Moreover, the use of cluster analysis is one of the tools for analyzing the stability
of topics [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] and the optimal number of topics [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ], but it does not consider the
bene ts of the special training capabilities of the topic model with sequential
regularization and dense representation of word-vectors.
      </p>
      <p>
        To validate the quality of clusters, quite a lot of metrics have been developed.
For example, Partition Coe cient [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], Dunn Index [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], as well as DPI [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and
its modi cations [
        <xref ref-type="bibr" rid="ref23 ref47">23, 47</xref>
        ], Silhouette [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ], which are involved in clustering
algorithms. Nevertheless, in the case of a topic model, we already get clusters of
topics and do not need a clustering algorithm, but only to evaluate the clusters
obtained. For validation of clusters it is necessary to consider them in space
with concepts of proximity and distance. For words, such a space is a vector
representation of words. Signi cant results in this direction were obtained in
researches [
        <xref ref-type="bibr" rid="ref11 ref38 ref46">38, 11, 46</xref>
        ]. Words presented in the form of dense vectors, re ect the
semantic representation and have the properties of proximity and distance.
Therefore, presenting the topics in the form of dense vectors, the authors created a
new variation of the DPI metric for the topics, which the authors called cDPI.
      </p>
      <p>The remainder of the paper is described as follows: the proposed methodology
and research hypothesis are presented in Section 2; the results of testing a new
quality metric are explained in Section 3. We conclude our paper in Section 4.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Research methodology</title>
      <p>Consider ways to build a topic model for a speci c collection of documents.
Collection is homogeneous if it contains documents of the same type. For example, a
collection of scienti c articles from one conference, created on a single template,
is homogeneous. In the case of a homogeneous collection of scienti c articles,
each document has a similar structure, postulated by a conference template. All
scienti c articles consist of introduction, presentation of research results and
conclusion. Thus, it is possible to present a document in the form of a distribution
of the main topic and auxiliary topics: introduction and conclusion.</p>
      <p>Of course, the main topics in di erent documents may be di erent. However,
the collection of scienti c articles may be limited to the choice of certain headings
from the thematic rubrics of the conference. Then the number of topics we will
know. Figure 1 shows matrix distribution of topics on the documents.</p>
      <p>
        As we see on the left side of Figure 1, topic model leads to the emphasis of
topics and their distribution homogeneously over the documents. Such a picture
of the probabilities of the \topics-documents" matrix can be obtained using, e.g.,
models based on the LDA algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In addition, the right side of Figure 1
shows the result of the model with sequential ARTM [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ]. The main and auxiliary
topics are highlighted through the management of the learning process of the
model. The principle of classifying a topic as auxiliary may be formulated as the
existence of such a topic in the overwhelming number of documents. That is,
the probabilities of the auxiliary topics will be distributed uniformly and tightly
across the documents. Furthermore, the main topic will be a sparse vector for
each document, since each document is characterized by one main topic.
      </p>
      <p>We show that the existing internal metrics of the topic model are not suitable
for determining the optimal number of topics. To do this, consider the internal
automated metrics of the quality of topics. We introduce the concept of core
topics:</p>
      <p>Wt = fw 2 W j p (tjw)
thresholdg :</p>
      <p>The following quality metrics of the topic model can be calculated based on
the topics kernel:</p>
      <p>Purity of the topics : P urity = Pw2Wt p(wjt)
Size of the topic kernel : jWtj
Contrast of the topics: jW1tj Pw2Wt p(tjw)
Coherence of the topics: Coht = k(k2 1) kiP=11 jP=k1 PMI(wi; wj ), where k is the
interval in which the combined use of words is calculated, point-wise mutual</p>
      <p>N Nwiwj , Nwiwj { the number of documents
information PMI(wi; wj ) = log Nwi Nwj
in which words wi and wj appear in interval k at least once. Nwj { the
number of documents in which the word wi appear at least once, and N is
the number of words in the dictionary.</p>
      <p>As can be seen from the formulas for the internal metrics of the topic model,
each of these metrics can be measured for a di erent number of topics (tn).
Consider the behavior of the metric Kernel size depending on the number of
topics. With an increase in the number of topics, the core size will decrease,
since the normalization conditions must be satis ed when constructing the
matrices \topics-words" and \documents-topics": the sum of the probabilities must
be equal to one. For metrics, the Purity of topics and the Contrast of topics,
the nature of changes with an increase in the number of topics will also be
monotonously decreasing, since the sum of the probabilities of the topics included
in the core will decrease. On the other hand, for the metric, Coherence to topics,
behavior with an increase in the number of topics will be monotonously
increasing, as the contribution from PMI will grow. The speci c nature of the changes
in the metrics examined may vary; therefore it is advisable to try to nd the
extreme point using numerical methods, if it is possible.</p>
      <p>
        The quality of the topics of short messages from the point of view of
clusters was reviewed in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] using NMF (Non-negative Matrix Factorization) and
metrics re ecting the entropy of clusters. The matrix approach (LSI + SVD) to
the selection of clusters of topics from the program code was investigated in [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]
with a modi ed vector proximity metric. The research of the topic model's
quality [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] use metric Silhouette Coe cient [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ] with Euclidean distance for sparse
subject vectors. Consequently, in these works, clusters in the space of dense
vectors{words constituting topics and non-Euclidean distances in metrics remain
unexplored.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref16 ref22 ref25">16, 25, 22</xref>
        ], the instability of topics with respect to the order of processed
documents was discovered and investigated. Therefore, to calculate the quality
metrics of the topics, it is necessary to perform calculations for the corpus of
documents with a random order to eliminate the dependence on the order of
documents. The possibility of stabilizing the topic model with the help of
regularization was shown in [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. Based on the analysis, the authors formulated a
methodological framework, depicted as a diagram in Figure 2.
For the experiment was used corpus of scienti c and technical articles on topics
related to the development of oil and gas elds. In total, 1695 articles in English
were selected in 10 areas of research according to the rubrics. The creation of a
dictionary for the selected corpus is described in detail in the previous study by
the authors [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. To build a topic model, the BigARTM library was used, which
allows for customization of the topic model by sequential regularization. The
choice and adjustment of the regularization parameters of the topic model were
made by the authors in a previous study [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. To transform the sparse space of
the vectors-words that make up the topics, the GloVe library was chosen [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ].
To obtain a visual representation of the form of a dense representation of topics,
a projection was made on a two-dimensional space with the distances preserved
using the MDS library [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Figure 3 presents the view of obtained clusters of
topics.
      </p>
      <p>In Figure 3, two-dimensional projections of words from topics are highlighted
with di erent markers. Ovals emphasize precise visual grouping of words in the
topics.</p>
      <p>As we can see from Figure 4, the nature of the dependencies is monotonous
and does not allow to determine the optimal number of topics. Measurements
of the main internal metrics are made for 1000 di erent random orders of
documents. The y {axis represents the value of one standard deviation. Evidently that
for the metric the Contrast of the core, the deviations are minimal. For metrics,
Purity and Coherence of the core the greater values characterize the best quality
of the topic model.</p>
      <p>
        A characteristic point can be considered the number of topics equal to 12,
when the curves of changes in the metric Purity and Coherence of topics intersect.
Consider the dependencies of the following metrics: Calinski-Harabaz Index [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ],
Silhouette Coe cient [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ], used to validate the number of clusters.
      </p>
      <p>According to Figure 5, the Calinski-Harabaz Index and Silhouette Coe cient
metrics do not make it possible to determine the optimal number of topics.
As the number of topics increases, the values of these metrics decrease, which
means that clusters become worse from the point of view of these metrics. The
cDBI metric developed by the authors and shown in Figure 6 behaves di erently
depending on the number of topics.</p>
      <p>
        In Figure 6 maximum clearly expressed with the number of topics equal to
16. The algorithm for calculating the cDBI metric is based on the ideology of
the Davies Bouldin Index metric proposed in [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and modi ed in [
        <xref ref-type="bibr" rid="ref23 ref47">23, 47</xref>
        ].
Result: cDBI
V := GloV e(ART M (tn; ; (corpus of texts))
for t 2 W : do
      </p>
      <p>Ct := Pi2t Vt(i)
Dt := dim1 t P</p>
      <p>Ct Vt((ii))
i2t jCtj jVt j
end
cDBI := dim1 W P</p>
      <p>Dt
t2T Ct</p>
      <p>Algorithm 1: Calculation of cDBI metrics.</p>
      <p>In the above Algorithm 1 T denotes the number of selected, { this
regularizing coe cients. Thus, using the cDBI metric, it is possible to nd the optimal
number of topics for a collection of documents.
4</p>
    </sec>
    <sec id="sec-3">
      <title>Conclusions</title>
      <p>The authors investigated the question of choosing the optimal number of topics
for building a topic model for a given corpus of texts. The result of this study
was a technique that allows you to determine the optimal number of topics for
corpus of texts.</p>
      <p>It should be said that the proposed method was experimentally con rmed
under the following conditions:
{ A small collection of documents;
{ English language of documents (monolingual);
{ Thematic uniformity.</p>
      <p>
        An important methodological trick of the authors is the preparation of a
topic model using sequential regularization. In previous studies of this collection
of documents [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], the authors obtained numerical estimates of the coe cients
for the regularizing components of the topic model ( ).
      </p>
      <p>When forming a collection of texts, conditions were set that limited the
number of topics of scienti c articles according to the topic rubrics to 10. The
essence of the experiment was to con rm the selected number of topics using an
optimization approach based on the quality metric developed by the authors of
the topic model | cDBI.</p>
      <p>As a result, the experiment showed that the maximum value of the cDBI
metric for test corpus is achieved with the average number of topics equal to 16
with standard deviation 2. The result was obtained with a large number of model
training to eliminate the in uence of the order of documents in the collection.</p>
      <p>In conclusion, it is important to emphasize that this study can serve as a
methodological groundwork for the creation of software frameworks and proposes
support for solving one of the fundamental problems of semantic text processing:
determining the sense of a text fragment (article).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Agrawal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Menzies</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>What is wrong with topic modeling? and how to x it using search-based software engineering</article-title>
          .
          <source>Information and Software Technology</source>
          <volume>98</volume>
          ,
          <issue>74</issue>
          {88 (jun
          <year>2018</year>
          ). https://doi.org/10.1016/j.infsof.
          <year>2018</year>
          .
          <volume>02</volume>
          .005, https://doi.org/10.1016/j.infsof.
          <year>2018</year>
          .
          <volume>02</volume>
          .005
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Asuncion</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Teh</surname>
            ,
            <given-names>Y.W.</given-names>
          </string-name>
          :
          <article-title>On smoothing and inference for topic models</article-title>
          .
          <source>In: Proceedings of the Twenty-Fifth Conference on Uncertainty in Arti cial Intelligence</source>
          . pp.
          <volume>27</volume>
          {
          <fpage>34</fpage>
          . UAI '09, AUAI Press, Arlington, Virginia, United
          <string-name>
            <surname>States</surname>
          </string-name>
          (
          <year>2009</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>1795114</volume>
          .
          <fpage>1795118</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Batmanghelich</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saeedi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Narasimhan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gershman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Nonparametric spherical topic modeling with word embeddings</article-title>
          .
          <source>In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>2</volume>
          : Short Papers).
          <source>Association for Computational Linguistics</source>
          (
          <year>2016</year>
          ). https://doi.org/10.18653/v1/p16-2087, https://doi.org/10.18653/v1/p16-
          <fpage>2087</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bezdek</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          :
          <article-title>Cluster validity with fuzzy sets</article-title>
          .
          <source>Journal of Cybernetics</source>
          <volume>3</volume>
          (
          <issue>3</issue>
          ),
          <volume>58</volume>
          {
          <fpage>73</fpage>
          (
          <year>1973</year>
          ). https://doi.org/10.1080/01969727308546047, https://doi.org/10.1080/01969727308546047
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bicalho</surname>
          </string-name>
          , P.V.,
          <string-name>
            <surname>d</surname>
          </string-name>
          . O.
          <string-name>
            <surname>Cunha</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mourao</surname>
            ,
            <given-names>F.H.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pappa</surname>
            ,
            <given-names>G.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meira</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Generating cohesive semantic topics from latent factors</article-title>
          .
          <source>In: 2014 Brazilian Conference on Intelligent Systems</source>
          . pp.
          <volume>271</volume>
          {
          <fpage>276</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (Oct
          <year>2014</year>
          ). https://doi.org/10.1109/bracis.
          <year>2014</year>
          .
          <volume>56</volume>
          , https://doi.org/10.1109/bracis.
          <year>2014</year>
          .56
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Bing</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bunea</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wegkamp</surname>
            ,
            <given-names>M.H.:</given-names>
          </string-name>
          <article-title>A fast algorithm with minimax optimal guarantees for topic models with an unknown number of topics</article-title>
          . CoRR abs/
          <year>1805</year>
          .
          <volume>06837</volume>
          (
          <issue>05</issue>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Binkley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heinz</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lawrie</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Overfelt</surname>
          </string-name>
          , J.:
          <article-title>Understanding lda in source code analysis</article-title>
          .
          <source>In: Proceedings of the 22Nd International Conference on Program Comprehension</source>
          . pp.
          <volume>26</volume>
          {
          <fpage>36</fpage>
          .
          <source>ICPC</source>
          <year>2014</year>
          , ACM, New York, NY, USA (
          <year>2014</year>
          ). https://doi.org/10.1145/2597008.2597150, http://doi.acm.
          <source>org/10</source>
          .1145/2597008.2597150
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          , Gri ths, T.L.,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.:</given-names>
          </string-name>
          <article-title>The nested chinese restaurant process and bayesian nonparametric inference of topic hierarchies</article-title>
          .
          <source>J. ACM</source>
          <volume>57</volume>
          (
          <issue>2</issue>
          ), 7:
          <issue>1</issue>
          {7:30 (feb
          <year>2010</year>
          ). https://doi.org/10.1145/1667053.1667056, http://doi.acm.
          <source>org/10</source>
          .1145/1667053.1667056
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.</given-names>
          </string-name>
          , Gri ths, T.L.,
          <string-name>
            <surname>Tenenbaum</surname>
            ,
            <given-names>J.B.</given-names>
          </string-name>
          :
          <article-title>Hierarchical topic models and the nested chinese restaurant process</article-title>
          .
          <source>In: Proceedings of the 16th International Conference on Neural Information Processing Systems</source>
          . pp.
          <volume>17</volume>
          {
          <fpage>24</fpage>
          . NIPS'03, MIT Press, Cambridge, MA, USA (
          <year>2003</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2981345</volume>
          .
          <fpage>2981348</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.</given-names>
          </string-name>
          :
          <article-title>Latent dirichlet allocation</article-title>
          .
          <source>J. Mach. Learn. Res</source>
          .
          <volume>3</volume>
          ,
          <issue>993</issue>
          {1022 (mar
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>5</volume>
          ,
          <issue>135</issue>
          {
          <fpage>146</fpage>
          (
          <year>2017</year>
          ), http://aclweb.org/anthology/Q17-1010
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Borg</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Groenen</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Modern multidimensional scaling: Theory and applications</article-title>
          .
          <source>Journal of Educational Measurement</source>
          <volume>40</volume>
          (
          <issue>3</issue>
          ),
          <volume>277</volume>
          { 280 (sep
          <year>2003</year>
          ). https://doi.org/10.1111/j.1745-
          <fpage>3984</fpage>
          .
          <year>2003</year>
          .tb01108.x, https://doi.org/10.1111/j.1745-
          <fpage>3984</fpage>
          .
          <year>2003</year>
          .tb01108.x
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Bryant</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sudderth</surname>
            ,
            <given-names>E.B.</given-names>
          </string-name>
          :
          <article-title>Truly nonparametric online variational inference for hierarchical dirichlet processes</article-title>
          .
          <source>In: Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 2</source>
          . pp.
          <volume>2699</volume>
          {
          <fpage>2707</fpage>
          . NIPS'
          <volume>12</volume>
          , Curran Associates Inc.,
          <source>USA</source>
          (
          <year>2012</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2999325</volume>
          .
          <fpage>2999436</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Calinski</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harabasz</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A dendrite method for cluster analysis</article-title>
          .
          <source>Communications in Statistics 3(1)</source>
          ,
          <volume>1</volume>
          {
          <fpage>27</fpage>
          (
          <year>1974</year>
          ). https://doi.org/10.1080/03610927408827101, https://www.tandfonline.com/doi/abs/10.1080/03610927408827101
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boyd-Graber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gerrish</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          :
          <article-title>Reading tea leaves: How humans interpret topic models</article-title>
          .
          <source>In: Proceedings of the 22Nd International Conference on Neural Information Processing Systems</source>
          . pp.
          <volume>288</volume>
          {
          <fpage>296</fpage>
          . NIPS'
          <volume>09</volume>
          , Curran Associates Inc.,
          <source>USA</source>
          (
          <year>2009</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2984093</volume>
          .
          <fpage>2984126</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Chuang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roberts</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stewart</surname>
            ,
            <given-names>B.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tingley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grimmer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heer</surname>
          </string-name>
          , J.: Topiccheck:
          <article-title>Interactive alignment for assessing topic model stability</article-title>
          .
          <source>In: Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>175</volume>
          {
          <fpage>184</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2015</year>
          ). https://doi.org/10.3115/v1/
          <fpage>N15</fpage>
          -1018, http://aclweb.org/anthology/N15-1018
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Das</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaheer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Gaussian lda for topic models with word embeddings</article-title>
          .
          <source>In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)</source>
          . pp.
          <volume>795</volume>
          {
          <fpage>804</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2015</year>
          ). https://doi.org/10.3115/v1/
          <fpage>P15</fpage>
          -1077, http://aclweb.org/anthology/P15-1077
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Davies</surname>
            ,
            <given-names>D.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bouldin</surname>
            ,
            <given-names>D.W.:</given-names>
          </string-name>
          <article-title>A cluster separation measure</article-title>
          .
          <source>IEEE Trans. Pattern Anal. Mach. Intell</source>
          .
          <volume>1</volume>
          (
          <issue>2</issue>
          ),
          <volume>224</volume>
          { 227 (feb
          <year>1979</year>
          ). https://doi.org/10.1109/TPAMI.
          <year>1979</year>
          .
          <volume>4766909</volume>
          , http://dx.doi.org/10.1109/TPAMI.
          <year>1979</year>
          .4766909
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19. Dunn†,
          <string-name>
            <surname>J.C.</surname>
          </string-name>
          :
          <article-title>Well-separated clusters and optimal fuzzy partitions</article-title>
          .
          <source>Journal of Cybernetics</source>
          <volume>4</volume>
          (
          <issue>1</issue>
          ),
          <volume>95</volume>
          {104 (jan
          <year>1974</year>
          ). https://doi.org/10.1080/01969727408546059, https://doi.org/10.1080/01969727408546059
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>El-Assady</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sevastjanova</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sperrle</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keim</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Collins,
          <string-name>
            <surname>C.</surname>
          </string-name>
          :
          <article-title>Progressive learning of topic modeling parameters: A visual analytics framework</article-title>
          .
          <source>IEEE Transactions on Visualization and Computer Graphics</source>
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <volume>382</volume>
          {391 (Jan
          <year>2018</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2017</year>
          .2745080
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Fang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Discovering research topics from library electronic references using latent dirichlet allocation</article-title>
          .
          <source>Library Hi Tech</source>
          <volume>36</volume>
          (
          <issue>3</issue>
          ),
          <volume>400</volume>
          {
          <volume>410</volume>
          (02
          <year>2018</year>
          ). https://doi.org/10.1108/LHT-06-2017-0132, https://app.dimensions.ai/details/publication/pub.1101114990
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Greene</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>O</given-names>
            <surname>'Callaghan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Cunningham</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>How many topics? stability analysis for topic models</article-title>
          .
          <source>In: Machine Learning and Knowledge Discovery in Databases</source>
          , pp.
          <volume>498</volume>
          {
          <fpage>513</fpage>
          . Springer: Berlin, Heidelberg (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Halkidi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Batistakis</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vazirgiannis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Clustering validity checking methods: Part ii</article-title>
          .
          <source>SIGMOD Rec</source>
          .
          <volume>31</volume>
          (
          <issue>3</issue>
          ),
          <volume>19</volume>
          {27 (sep
          <year>2002</year>
          ). https://doi.org/10.1145/601858.601862, http://doi.acm.
          <source>org/10</source>
          .1145/601858.601862
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Koltcov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Application of renyi and tsallis entropies to topic modeling optimization</article-title>
          .
          <source>Physica A: Statistical Mechanics and its Applications</source>
          <volume>512</volume>
          , 1192{1204 (dec
          <year>2018</year>
          ). https://doi.org/10.1016/j.physa.
          <year>2018</year>
          .
          <volume>08</volume>
          .050, https://doi.org/10.1016/j.physa.
          <year>2018</year>
          .
          <volume>08</volume>
          .050
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Koltcov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koltsova</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nikolenko</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Latent dirichlet allocation: Stability and applications to studies of user-generated content</article-title>
          .
          <source>In: Proceedings of the 2014 ACM Conference on Web Science</source>
          . pp.
          <volume>161</volume>
          {
          <fpage>165</fpage>
          . WebSci '14,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2014</year>
          ). https://doi.org/10.1145/2615569.2615680, http://doi.acm.
          <source>org/10</source>
          .1145/2615569.2615680
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Koltcov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nikolenko</surname>
            ,
            <given-names>S.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koltsova</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Filippov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bodrunova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Stable topic modeling with local density regularization</article-title>
          .
          <source>In: Internet Science</source>
          , pp.
          <volume>176</volume>
          {
          <fpage>188</fpage>
          . Springer International Publishing (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Koltsov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pashakhin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dokuka</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A full-cycle methodology for news topic modeling and user feedback research</article-title>
          . In: Staab,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Koltsova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            ,
            <surname>Ignatov</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.I</surname>
          </string-name>
          . (eds.) Social Informatics. pp.
          <volume>308</volume>
          {
          <fpage>321</fpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Krasnov</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ushmaev</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Exploration of hidden research directions in oil and gas industry via full text analysis of onepetro digital library</article-title>
          .
          <source>International Journal of Open Information Technologies</source>
          <volume>6</volume>
          (
          <issue>5</issue>
          ),
          <volume>7</volume>
          {
          <fpage>14</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Kuhn</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ducasse</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , G^rba, T.:
          <article-title>Semantic clustering: Identifying topics in source code</article-title>
          .
          <source>Information and Software Technology</source>
          <volume>49</volume>
          (
          <issue>3</issue>
          ),
          <volume>230</volume>
          {243 (mar
          <year>2007</year>
          ). https://doi.org/10.1016/j.infsof.
          <year>2006</year>
          .
          <volume>10</volume>
          .017, https://doi.org/10.1016/j.infsof.
          <year>2006</year>
          .
          <volume>10</volume>
          .017
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Law</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhuo</surname>
            ,
            <given-names>H.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rong</surname>
          </string-name>
          , E.: LTSG:
          <article-title>Latent topical skip-gram for mutually improving topic model and vector representations</article-title>
          .
          <source>In: Pattern Recognition and Computer Vision</source>
          , pp.
          <volume>375</volume>
          {
          <fpage>387</fpage>
          . Springer International Publishing (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Lipton</surname>
            ,
            <given-names>Z.C.</given-names>
          </string-name>
          :
          <article-title>The mythos of model interpretability</article-title>
          .
          <source>Queue</source>
          <volume>16</volume>
          (
          <issue>3</issue>
          ),
          <volume>30</volume>
          :
          <fpage>31</fpage>
          {
          <fpage>30</fpage>
          :57 (jun
          <year>2018</year>
          ). https://doi.org/10.1145/3236386.3241340, http://doi.acm.
          <source>org/10</source>
          .1145/3236386.3241340
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Mantyla</surname>
            ,
            <given-names>M.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Claes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farooq</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Measuring lda topic stability from clusters of replicated runs</article-title>
          .
          <source>In: Proceedings of the 12th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement</source>
          . pp.
          <volume>49</volume>
          :
          <issue>1</issue>
          {
          <issue>49</issue>
          :
          <article-title>4</article-title>
          . ESEM '18,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2018</year>
          ). https://doi.org/10.1145/3239235.3267435, http://doi.acm.
          <source>org/10</source>
          .1145/3239235.3267435
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Mehta</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caceres</surname>
            ,
            <given-names>R.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carter</surname>
            ,
            <given-names>K.M.:</given-names>
          </string-name>
          <article-title>Evaluating topic quality using model clustering</article-title>
          .
          <source>In: 2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM)</source>
          . pp.
          <volume>178</volume>
          {
          <issue>185</issue>
          (Dec
          <year>2014</year>
          ). https://doi.org/10.1109/cidm.
          <year>2014</year>
          .
          <volume>7008665</volume>
          , https://doi.org/10.1109/cidm.
          <year>2014</year>
          .7008665
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Mimno</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Bayesian checking for topic models</article-title>
          .
          <source>In: Proceedings of the Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>227</volume>
          {
          <fpage>237</fpage>
          . EMNLP '
          <volume>11</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computational Linguistics, Stroudsburg, PA, USA (
          <year>2011</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2145432</volume>
          .
          <fpage>2145459</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Newman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lau</surname>
            ,
            <given-names>J.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grieser</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baldwin</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Automatic evaluation of topic coherence</article-title>
          . In: Human Language Technologies:
          <article-title>The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics</article-title>
          . pp.
          <volume>100</volume>
          {
          <fpage>108</fpage>
          . HLT '
          <volume>10</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computational Linguistics, Stroudsburg, PA, USA (
          <year>2010</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>1857999</volume>
          .
          <fpage>1858011</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>D.Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Billingsley</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          , M.:
          <article-title>Improving topic models with latent feature word representations</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>3</volume>
          ,
          <issue>299</issue>
          {
          <fpage>313</fpage>
          (
          <year>2015</year>
          ), http://aclweb.org/anthology/Q15-1022
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Nikolenko</surname>
            ,
            <given-names>S.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koltcov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koltsova</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Topic modelling for qualitative studies</article-title>
          .
          <source>Journal of Information Science</source>
          <volume>43</volume>
          (
          <issue>1</issue>
          ),
          <volume>88</volume>
          {102 (jul
          <year>2016</year>
          ). https://doi.org/10.1177/0165551515617393, https://doi.org/10.1177/0165551515617393
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <fpage>1543</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2014</year>
          ). https://doi.org/10.3115/v1/
          <fpage>D14</fpage>
          -1162, http://aclweb.org/anthology/D14-1162
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Rossetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stella</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zanker</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Towards explaining latent factors with topic models in collaborative recommender systems</article-title>
          .
          <source>In: 2013 24th International Workshop on Database and Expert Systems Applications. IEEE</source>
          (09
          <year>2013</year>
          ). https://doi.org/10.1109/DEXA.
          <year>2013</year>
          .
          <volume>26</volume>
          , https://doi.org/10.1109/dexa.
          <year>2013</year>
          .26
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Silhouettes: A graphical aid to the interpretation and validation of cluster analysis</article-title>
          .
          <source>J. Comput. Appl</source>
          . Math.
          <volume>20</volume>
          (
          <issue>1</issue>
          ),
          <volume>53</volume>
          {65 (nov
          <year>1987</year>
          ). https://doi.org/10.1016/
          <fpage>0377</fpage>
          -
          <lpage>0427</lpage>
          (
          <issue>87</issue>
          )
          <fpage>90125</fpage>
          -
          <lpage>7</lpage>
          , http://dx.doi.org/10.1016/
          <fpage>0377</fpage>
          -
          <lpage>0427</lpage>
          (
          <issue>87</issue>
          )
          <fpage>90125</fpage>
          -
          <lpage>7</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41.
          <string-name>
            <surname>Seroussi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bohnert</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zukerman</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Authorship attribution with author-aware topic models</article-title>
          .
          <source>In: Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers - Volume 2</source>
          . pp.
          <volume>264</volume>
          {
          <fpage>269</fpage>
          . ACL '
          <volume>12</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computational Linguistics, Stroudsburg, PA, USA (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Storn</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Price</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Di erential evolution { a simple and e cient heuristic for global optimization over continuous spaces</article-title>
          .
          <source>J. of Global Optimization</source>
          <volume>11</volume>
          (
          <issue>4</issue>
          ),
          <volume>341</volume>
          {359 (dec
          <year>1997</year>
          ). https://doi.org/10.1023/A:1008202821328, https://doi.org/10.1023/A:1008202821328
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          43.
          <string-name>
            <surname>Teh</surname>
            ,
            <given-names>Y.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beal</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.:</given-names>
          </string-name>
          <article-title>Sharing clusters among related groups: Hierarchical dirichlet processes</article-title>
          .
          <source>In: Proceedings of the 17th International Conference on Neural Information Processing Systems</source>
          . pp.
          <volume>1385</volume>
          {
          <fpage>1392</fpage>
          . NIPS'04, MIT Press, Cambridge, MA, USA (
          <year>2004</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2976040</volume>
          .
          <fpage>2976214</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          44.
          <string-name>
            <surname>Vorontsov</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potapenko</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Plavin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Additive regularization of topic models for topic selection and sparse factorization</article-title>
          .
          <source>In: Statistical Learning and Data Sciences</source>
          , pp.
          <volume>193</volume>
          {
          <fpage>202</fpage>
          . Springer International Publishing (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          45.
          <string-name>
            <surname>Wallach</surname>
            ,
            <given-names>H.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murray</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salakhutdinov</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mimno</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Evaluation methods for topic models</article-title>
          .
          <source>In: Proceedings of the 26th Annual International Conference on Machine Learning</source>
          . pp.
          <volume>1105</volume>
          {
          <fpage>1112</fpage>
          . ICML '09,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2009</year>
          ). https://doi.org/10.1145/1553374.1553515, http://doi.acm.
          <source>org/10</source>
          .1145/1553374.1553515
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          46.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>L.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fisch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chopra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adams</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bordes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
          </string-name>
          , J.: Starspace:
          <article-title>Embed all the things! In: AAAI (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          47.
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>X.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beni</surname>
          </string-name>
          , G.:
          <article-title>A validity measure for fuzzy clustering</article-title>
          .
          <source>IEEE Trans. Pattern Anal. Mach. Intell</source>
          .
          <volume>13</volume>
          (
          <issue>8</issue>
          ),
          <volume>841</volume>
          {847 (aug
          <year>1991</year>
          ). https://doi.org/10.1109/34.85677, http://dx.doi.org/10.1109/34.85677
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>