<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A QUBO Formulation of the k-Medoids Problem</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>C. Bauckhage</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>N. Piatkowski</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>R. Sifa</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>D. Hecker</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. Wrobel</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AI Group, TU Dortmund</institution>
          ,
          <addr-line>Dortmund</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>B-IT, University of Bonn</institution>
          ,
          <addr-line>Bonn</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Fraunhofer Center for Machine Learning</institution>
          ,
          <addr-line>Sankt Augustin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Fraunhofer IAIS</institution>
          ,
          <addr-line>Sankt Augustin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We are concerned with k-medoids clustering and propose a quadratic unconstrained binary optimization (QUBO) formulation of the problem of identifying k medoids among n data points without having to cluster the data. Given our QUBO formulation of this NP-hard problem, it should be possible to solve it on adiabatic quantum computers. Copyright c 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Quadratic unconstrained binary optimization problems (QUBOs) are concerned
with nding an n-dimensional binary vector that minimizes a quadratic objective
z = argmin z|Qz + z|q:
z2f0;1gn
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
They thus pose combinatorial optimization problems which generally prove to
be NP-hard. Indeed, di cult problems such as capital budgeting or task
allocation in operations research, constraint satisfaction in AI, or maximum diversity,
maximum clique, or graph partitioning in data mining and machine learning are
but a few examples of where QUBOs occur in practice [1].
      </p>
      <p>Owing to their practical importance, there exists a venerable literature on
QUBOs and e cient solution strategies (for an extensive survey, see [2] and the
references therein). More recently, however, interest in QUBOs has also arisen
in a di erent context. Since adiabatic quantum computers such as produced by
D-Wave [3,4] are designed to solve them, research on QUBO reformulations of
various problems has noticeably intensi ed. Indeed, Boolean satis ability [5],
graph cuts [6,7,8], graph isomorphisms [9], binary clustering [6,10], or classi er
training [11,12] have lately been solved via adiabatic quantum computing.</p>
      <p>Following this line of research, we propose a QUBO formulation of the
problem of identifying k medoids among n data points which, to our knowledge,
has neither been attempted nor reported before. While we do not consider a
quantum computing implementation itself, our result indicates that prototype
extraction can be accomplished on quantum computers.</p>
      <p>mean
medoid
mean
medoid
(a) Gaussian blob
(b) ring
mean
medoid
(c) spiral</p>
      <p>Next, we brie y summarize the notion of a medoid and the ideas behind
k-medoids clustering. We then present our QUBO formulation of the k-medoids
problem and discuss simple examples that illustrate the practical performance of
our approach. Finally, we review related work and summarize our contributions.
2</p>
      <p>
        k-Medoids Clustering
Consider a sample X = fx1; : : : ; xng of n Euclidean data points xi 2 Rm. The
well established sample mean
= argmin X
x2Rm
xi2X
xi
x 2 = n1 X xi
xi2X
is a frequently used summary statistic for such data. The so called sample medoid
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
m = argmin X
xj2X xi2X
xi
xj
2
on the other hand, is arguably less widely known but has interesting applications
as well.
      </p>
      <p>
        Figure 1 illustrates characteristics of means and medoids. Both minimize
sums of distances to available sample points. However, while the mean is not
necessarily contained in the sample, the medoid always coincides with an element
of the sample and, for squared Euclidean distances as in (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ), it is easy to prove
that this element is the sample point closest to the mean [13].
      </p>
      <p>
        An interesting feature of medoids is that they result from evaluating squared
distances between given data only.1 As these can be precomputed and stored in
an n n distance matrix D where Dij = d2(xi; xj ), medoids can be computed
1 Note once again that x in equation (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) may not be contained in X whereas xi and
xj in equation (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) always are.
      </p>
      <p>M = fj1; j2; : : : ; jkg</p>
      <p>I
repeat
for l = 1; : : : ; k do
determine cluster</p>
      <p>Cl = i 2 I</p>
      <sec id="sec-1-1">
        <title>Dijl</title>
      </sec>
      <sec id="sec-1-2">
        <title>Dijq</title>
        <p>for l = 1; : : : ; k do
update medoid index
jl = argmin X
j2Cl i2Cl</p>
        <p>
          Dij
until clusters stabilize
(
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
(
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
Algorithm 1 k-medoids clustering via Lloyd's algorithm
Require: index set I = f1; : : : ; ng, distance matrix D 2 Rn n, and parameter k 2 N
initialize set of k cluster medoid indices
from relational data. Contrary to means, medoids may thus also be estimated
on sets of strings or graphs or other non-numeric data as long as there is an
appropriate distance measure d( ; ) [13].
        </p>
        <p>Moreover, since medoids coincide with actual data points, analysts often
nd them easy to interpret. This is why k-medoids clustering is an increasingly
popular tool whenever explainable or physically plausible prototypes need to be
determined [14,15,16,17,18].</p>
        <p>Conceptually, k-medoids clustering is almost identical to k-means clustering.
To accomplish it, we may, for instance, simply adapt Lloyd's algorithm [19]. In
other words, given randomly initialized medoids, we may determine clusters by
assigning data points to their closest medoid, update the medoid of each cluster,
and repeat these steps until clusters stabilize.</p>
        <p>
          Note, however, that (local) medoids, too, are relational objects that can solely
be determined from analyzing a distance matrix. Given such a distance matrix
for a data set X , k-medoids clustering can thus be performed on index sets only.
We demonstrate this in Algorithm 1 where I = f1; 2; : : : ; ng indexes the data
in X . Having initialized a set M = fj1; j2; : : : ; jkg I of medoid indices, each
cluster Cl is an index set containing the indices of those data points xi that are
closest to medoid mjl (equation (
          <xref ref-type="bibr" rid="ref4">4</xref>
          )). Given the clusters, each medoid index jl is
then set to the index of the medoid of the points indexed by Cl (equation (
          <xref ref-type="bibr" rid="ref5">5</xref>
          )).
        </p>
        <p>Similar to k-means clustering, k-medoids clustering subdivides the input
space into convex cells each of which is de ned w.r.t. a prototype. Just as in
k-means clustering via MacQueen's algorithm [20], k-medoids clustering could
therefore in principle be performed in two steps where medoids are determined
rst and data are assigned to them second. However, contrary to k-means
clustering, the combinatorial nature of k-medoids clustering, which selects prototypes
from a discrete set of data points, prevents the use of continuous MacQueen-type
updates of prototypes.</p>
        <p>Indeed, we are not aware of prior work on how to nd local medoids without
having to compute clusters. The quadratic unconstrained binary optimization
formulation of the problem which we present in the next section, however, can
accomplish this.
3</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>A QUBO Formulation for k-Medoids Estimation</title>
      <p>Our idea for how to select k local medoids among n data points is based on an
observation regarding classical k-means clustering.</p>
      <p>Recall that, if a set X of n data points is partitioned into k clusters Cl of nl
elements whose mean is l, adding the within cluster scatter
and the between cluster scatter</p>
      <p>k
SW = X X x
l=1 x2Cl
l</p>
      <p>2
SB =
1 Xk Xk ni nj
2 n i=1 j=1
i
j
2
yields a constant; that is, we have SW + SB = c.</p>
      <p>This follows from Fisher's analysis of variance [21,22] and is to say that the
two scatter values are inverse under addition: either SW is large and SB is small,
or SW is small and SB is large. Since a small within cluster scatter is usually
taken to be the objective of k-means clustering, we therefore observe that \good"
cluster means are close to the data points within their cluster and, at the same
time, far apart from each other.</p>
      <p>Consequently, we posit that the same should hold true for the medoids that
are supposed to result from k-medoids clustering.</p>
      <p>
        Note, however, that (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) involves variances about local centroids. Since medoids
are not necessarily central, we henceforth work with a more robust measure of
similarity. In particular, we consider Welsch's M -estimator
ij = 1
exp
21 Dij
which is known from robust regression [23] and also referred to as the correntropy
loss [24,25].
3.1
      </p>
      <p>
        A QUBO to identify far apart data points
The problem of selecting k mutually far apart objects among a total of n objects
is known as the max-sum dispersion problem [26] and can be formalized as
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
(
        <xref ref-type="bibr" rid="ref7">7</xref>
        )
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
M
= argmax
      </p>
      <p>M I
s:t:</p>
      <p>Looking at this problem, we note that, upon introducing binary indicator
vectors z 2 f0; 1gn whose entries are given by
it can also be written in terms of a constrained binary optimization problem
follows: given an n n similarity matrix
I = f1; 2; : : : ; ng, determine a subset M
over n objects which are indexed by</p>
      <p>I, jM j = k such that
z = za2rgf0m;1agxn 21 z|
= argmax z| 1</p>
      <p>z2f0;1gn 2
where 1 2 Rn denotes the vector of all ones.</p>
      <p>
        If we further note that z|1 = k , (z|1 k)2 = 0, we nd that (
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
can equivalently be expressed as a quadratic unconstrained binary optimization
problem, namely
z = za2rgf0m;1agxn 21 z|
z
(z|1
k)2
where 2 R is a Lagrange multiplier. Treating this multiplier as a constant and
expanding the expression on the right hand side, we obtain
Using the above notation, the problem of selecting k most central objects among
a total of n objects can be formalized as follows
      </p>
      <p>M
= argmin</p>
      <p>M I
s:t:</p>
      <p>X X
i2M j2I
jMj = k:
ij</p>
      <p>
        Introducing binary indicator vectors and a Lagrange multiplier, reasoning as
above then leads to the following quadratic unconstrained binary optimization
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        )
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
(
        <xref ref-type="bibr" rid="ref12">12</xref>
        )
(
        <xref ref-type="bibr" rid="ref13">13</xref>
        )
(
        <xref ref-type="bibr" rid="ref14">14</xref>
        )
(
        <xref ref-type="bibr" rid="ref15">15</xref>
        )
(
        <xref ref-type="bibr" rid="ref16">16</xref>
        )
problem
z = argmin z|
      </p>
      <p>
        z2f0;1gn
In order to combine the QUBO that identi es far apart points (
        <xref ref-type="bibr" rid="ref15">15</xref>
        ) and the
QUBO that identi es central points (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ) into a single model that identi es rather
central points that are rather far apart, we adhere to common practice [27].
That is, we introduce two additional tradeo parameters ; 2 R to weigh the
contributions of either model. This way, we obtain
z = argmin z|
z2f0;1gn
11|
Working with Welsch's function in (
        <xref ref-type="bibr" rid="ref8">8</xref>
        ) has the additional bene t that it maps
squared distances such as kxi xj k2 into the interval [0; 1].
      </p>
      <p>
        For the weighted, model speci c contributions that occur in (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ), we therefore
have the following two upper bounds
21 z|
z|
z
1
12 k2 1
n k 1:
This suggests to set the weighting parameters to = 1=k and = 1=n so as
to normalize the two contribution to about the same range. For the Lagrange
multiplier which enforces solutions z to have k entries equal to 1, we choose
= 2 so as to prioritize this constraint.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Practical Examples</title>
      <p>
        Next, we present two admittedly simple experiments that illustrate the behavior
of the QUBOs in (
        <xref ref-type="bibr" rid="ref15">15</xref>
        ), (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ), and (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ). Note again that our primary goal in this
paper is to investigate whether it is possible to identifying k medoids among n
data points without having to cluster the data; e ciency and scalability are not
our main concerns. Correspondingly, we consider two samples of 2D data points
that are small enough to allow for brute force estimation of the minimizers of
the above QUBOs. In other words, in both experiments, we compute similarity
matrices over n 2 f12; 16g data points, chose , , and as above and then
evaluate each of the 2n possible solutions to the QUBOs in (
        <xref ref-type="bibr" rid="ref15">15</xref>
        ), (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ), and (
        <xref ref-type="bibr" rid="ref19">19</xref>
        )
in order to determine the respective best one.
      </p>
      <p>
        The data in our rst experiment were deliberately chosen such that extremal,
central, as well as locally medoidal data points can be easily identi ed by visual
(
        <xref ref-type="bibr" rid="ref17">17</xref>
        )
(
        <xref ref-type="bibr" rid="ref18">18</xref>
        )
(
        <xref ref-type="bibr" rid="ref20">20</xref>
        )
(
        <xref ref-type="bibr" rid="ref21">21</xref>
        )
(a) 2D data
(b) solution to (
        <xref ref-type="bibr" rid="ref15">15</xref>
        )
(c) solution to (
        <xref ref-type="bibr" rid="ref18">18</xref>
        )
(d) solution to (
        <xref ref-type="bibr" rid="ref19">19</xref>
        )
inspection. They can be seen in Fig. 2a which shows n = 12 two-dimensional data
points which form 4 apparent clusters. Setting k = 4 and solving the QUBOs
in (
        <xref ref-type="bibr" rid="ref15">15</xref>
        ), (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ), and (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ) produces indicator vectors that identify the data points
highlighted in Fig. 2b{2d. Looking at these results suggests that they are well
in line with what human analysts would deem reasonable.
      </p>
      <p>
        In our second experiment, we generated 100 sets of n = 16 data points
sampled from 3 bivariate Gaussian distributions; Fig. 3a shows an example.
Setting k = 3, we then solved the QUBOs in (
        <xref ref-type="bibr" rid="ref15">15</xref>
        ), (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ), and (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ). Exemplary
results can be seen in Fig. 3b{3d. The three extremal data points in Fig. 3b make
intuitive sense. The fact that the three rather central data points in Fig. 3c are
all situated in the cluster to the bottom left is due to the fact that this is the
largest of the three clusters; here, our use of Welsch's M -estimator has the e ect
that points which are central w.r.t. several nearby points are being favored. The
local medoids highlighted in Fig. 3d are again intuitive.
      </p>
      <p>
        For each of the 100 data sets considered in this experiment, we also compared
the k = 3 medoids obtained from solving (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ) to those produced by Algorithm 1.
In each of our 100 tests, we found them to be identical. While such a perfect
agreement between both methods is likely an artifact of the small sample sizes
we worked with, it nevertheless corroborates the utility of our QUBO based
approach to the k-medoids problem.
5
      </p>
    </sec>
    <sec id="sec-4">
      <title>Related Work</title>
      <p>As of this writing, there are three major algorithmic approaches to k-means
clustering, namely those due to Lloyd [19], MacQueen [20], and Hartigan and
Wong [28]. Lloyd- and Hartigan-type approaches alternatingly update means
and clusters and have been applied to the combinatorial problem of k-medoids
clustering, too. Examples for the former can be found in [13] and [29]; examples
of the latter include the algorithms PAM and CLARA as well as variations
thereof [30,31,32].</p>
      <p>
        However, to the best of our knowledge, MacQueen type procedures, which
determine local means without having to compute clusters, have not yet been
reported for the k-medoids setting. The QUBO presented in equation (
        <xref ref-type="bibr" rid="ref19">19</xref>
        ) lls
this gap. While it is not an iterative method such as MacQueen's procedure, it
nevertheless suggests that medoids, too, can be estimated without clustering.
      </p>
      <p>Our derivation of a QUBO for the k-medoids problem was motivated by the
quest for quantum computing solutions for relational or combinatorial clustering.
Here, most prior work we are aware of has focused on solutions for k = 2 [6,7,8].
Work on solutions for k &gt; 2 can be found in [33].</p>
      <p>However, the methods proposed there are similar in spirit to kernel k-means
clustering in that they operate on binary cluster membership indicator matrices
Z 2 f0; 1gk n. This is to say that they perform quantum clustering in a manner
that produces clusters but does not identify cluster prototypes. Our approach
in this paper, however, only requires binary indicator vectors z 2 f0; 1gn and is,
once again, able to identify 1 k n prototypes without having to compute
clusters.
6</p>
    </sec>
    <sec id="sec-5">
      <title>Summary</title>
      <p>In this paper, we have proposed a quadratic unconstrained binary optimization
(QUBO) formulation of the problem of identifying k local medoids within a
sample of n data points. The basic idea is to trade o measures of central and
extremal tendencies of individual data points. Just as conventional approaches
to k-medoids clustering, our solution works with relational data and therefore
applies to a wide range of practical settings. However, to the best of our
knowledge, our solution is the rst that is capable of extracting k local medoids without
having to compute clusters.</p>
      <p>This capability comes at a price. While conventional k-medoids clustering as
well as our QUBO formulation constitute NP-hard combinatorial problems, the
former can be solved (approximately) by means of greedy heuristics. Yet, for the
latter, such heuristics are hard to conceive.</p>
      <p>However, an aspect of our model that may soon turn into a viable advantage
is that it can be easily solved on quantum computers. While this paper did not
consider corresponding quantum computing implementations, they are the topic
of ongoing work and results will be reported once available.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Kochenberger</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glover</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A Uni ed Framework for Modeling and Solving Combinatorial Optimization Problems: A Tutorial</article-title>
          . In Hager, W.,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pardalos</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prokopyev</surname>
          </string-name>
          , O., eds.:
          <source>Multiscale Optimization Methods and Applications</source>
          . Volume
          <volume>82</volume>
          <source>of NOIA</source>
          . Springer (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Kochenberger</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hao</surname>
            ,
            <given-names>J.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glover</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>The Unconstrained Binary Quadratic Programming Problem: A Survey</article-title>
          .
          <source>J. of Combinatorial Optimization</source>
          <volume>28</volume>
          (
          <issue>1</issue>
          ) (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , et al.:
          <article-title>Quantum Annealing with Manufactured Spins</article-title>
          .
          <source>Nature</source>
          <volume>473</volume>
          (
          <issue>7346</issue>
          ) (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>D</given-names>
            <surname>-</surname>
          </string-name>
          <article-title>Wave press release: D-Wave announces D-Wave 2000Q quantum computer and rst system order</article-title>
          (
          <year>Jan 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Farhi</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldstone</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gutmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sipser</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Quantum Computation by Adiabatic Evolution</article-title>
          .
          <source>arXiv:quant-ph/0001106</source>
          (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Bauckhage</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brito</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cvejoski</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ojeda</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sifa</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wrobel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Ising Models for Binary Clustering via Adiabatic Quantum Computing</article-title>
          .
          <source>In: Proc. EMMCVPR</source>
          . Volume
          <volume>10746</volume>
          of LNCS., Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Ushijima-Mwesigwa</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Negre</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mniszewski</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Graph Partitioning Using Quantum Annealing on the D-Wave System</article-title>
          .
          <source>In: Proc. Int. Workshop on Post Moores Era Supercomputing</source>
          , ACM (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. Junger,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Lobe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Mutzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Reinelt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Rendl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Rinaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Stollenwerk</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          :
          <article-title>Performance of a Quantum Annealer for Ising Ground State Computations on Chimera Graphs</article-title>
          . arXiv:
          <year>1904</year>
          .
          <article-title>11965 [cs</article-title>
          .DS] (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Calude</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dinneen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hua</surname>
          </string-name>
          , R.:
          <article-title>QUBO Formulations for the Graph Isomorphism Problem and Related Problems</article-title>
          .
          <source>Theoretical Computer Science</source>
          <volume>701</volume>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Bauckhage</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ojeda</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sifa</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wrobel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Adiabatic Quantum Computing for Kernel k=2 Means Clustering</article-title>
          .
          <source>In: Proc. KDML-LWDA</source>
          . (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Pudenz</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lidar</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Quantum Adiabatic Machine Learning</article-title>
          .
          <source>Quantum Information Processing</source>
          <volume>12</volume>
          (
          <issue>5</issue>
          ) (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Adachi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Henderson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Application of Quantum Annealing to Training of Deep Neural Networks</article-title>
          .
          <source>arXiv:1510</source>
          .06356 [quant-ph
          <string-name>
            <surname>]</surname>
          </string-name>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Bauckhage</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>NumPy / SciPy Recipes for Data Science: k-Medoids Clustering</article-title>
          . researchgate.
          <source>net (Feb</source>
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Drachen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sifa</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thurau</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The Name in the Game: Patterns in Character Names</article-title>
          and
          <string-name>
            <given-names>Game</given-names>
            <surname>Tags</surname>
          </string-name>
          .
          <source>Entertainment Computing</source>
          <volume>5</volume>
          (
          <issue>1</issue>
          ) (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Bauckhage</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Drachen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sifa</surname>
          </string-name>
          , R.:
          <article-title>Clustering Game Behavior Data</article-title>
          .
          <source>IEEE Trans. on Computational Intelligence and AI in Games</source>
          <volume>7</volume>
          (
          <issue>3</issue>
          ) (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Caro</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aarva</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deringer</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Csanyi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Laurila</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Reactivity of Amorphous Carbon Surfaces: Rationalizing the Role of Structural Motifs in Functionalization Using Machine Learning</article-title>
          .
          <source>Chemistry of Materials</source>
          <volume>30</volume>
          (
          <issue>21</issue>
          ) (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Molina</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vergari</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Di Mauro</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Natarajan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Esposito</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kersting</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Mixed Sum-Product Networks: A Deep Architecture for Hybrid Domains</article-title>
          .
          <source>In: Proc. AAAI</source>
          . (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , et al.:
          <article-title>A Universal Probe Set for Targeted Sequencing of 353 Nuclear Genes from Any Flowering Plant Designed Using k-Medoids Clustering</article-title>
          .
          <source>Systematic Biology (12</source>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Lloyd</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Least Squares Quantization in PCM</article-title>
          .
          <source>IEEE Trans. Information Theory</source>
          <volume>28</volume>
          (
          <issue>2</issue>
          ) (
          <year>1982</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>MacQueen</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Some Methods for Classi cation and Analysis of Multivariate Observations</article-title>
          .
          <source>In: Proc. Berkeley Symp. on Mathematical Statistics and Probability</source>
          . (
          <year>1967</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21. Fisher, R.:
          <article-title>On the Probable Error of a Coe cient Correlation Deduced from a Small Sample</article-title>
          .
          <source>Metron</source>
          <volume>1</volume>
          (
          <year>1921</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Bauckhage</surname>
          </string-name>
          , C.
          <article-title>: k-Means and Fisher's Analysis of Variance. researchgate</article-title>
          .net (May
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Dennis</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Welsch</surname>
          </string-name>
          , R.:
          <article-title>Techniques for Nonlinear Least Squares and Robust Regression</article-title>
          .
          <source>Communications in Statistics { Simulation and Computation</source>
          <volume>7</volume>
          (
          <issue>4</issue>
          ) (
          <year>1978</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pokharel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Principe</surname>
          </string-name>
          , J.:
          <article-title>Correntropy: Properties and Applications in Non-Gaussian Signal Processing</article-title>
          .
          <source>IEEE Trans. on Signal Processing</source>
          <volume>55</volume>
          (
          <issue>11</issue>
          ) (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suykens</surname>
          </string-name>
          , J.:
          <article-title>Learning with the Maximum Correntropy Criterion Induced Losses for Regression</article-title>
          .
          <source>J. of Machine Learning Research</source>
          <volume>16</volume>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Ravi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosenkrantz</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tayi</surname>
          </string-name>
          , G.:
          <article-title>Heuristic and Special Case Algorithms for Dispersion Problems</article-title>
          .
          <source>Operations Research</source>
          <volume>42</volume>
          (
          <issue>2</issue>
          ) (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Lucas</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Ising Formulations of Many NP Problems</article-title>
          .
          <source>Frontiers in Physics 2</source>
          (
          <issue>5</issue>
          ) (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Hartigan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <string-name>
            <surname>Algorithm</surname>
            <given-names>AS</given-names>
          </string-name>
          136:
          <article-title>A k-Means Clustering Algorithm</article-title>
          .
          <source>J. of the Royal Statistical Society C</source>
          <volume>28</volume>
          (
          <issue>1</issue>
          ) (
          <year>1979</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Park</surname>
            ,
            <given-names>H.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jun</surname>
            ,
            <given-names>C.H.</given-names>
          </string-name>
          :
          <article-title>A Simple and Fast Algorithm for K-Medoids Clustering</article-title>
          .
          <source>Expert Systems with Applications</source>
          <volume>36</volume>
          (
          <issue>2</issue>
          ) (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Kaufman</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Partitioning Around Medoids (Program PAM)</article-title>
          .
          <article-title>In: Finding Groups in Data: An Introduction to Cluster Analysis</article-title>
          . John Wiley &amp; Sons (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Kaufman</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Clustering Large Applications (Program CLARA)</article-title>
          .
          <article-title>In: Finding Groups in Data: An Introduction to Cluster Analysis</article-title>
          . John Wiley &amp; Sons (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Schubert</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Faster k-Medoids Clustering: Improving the PAM, CLARA, and CLARANS Algorithms</article-title>
          . arXiv:
          <year>1810</year>
          .
          <article-title>05691 [cs</article-title>
          .LG] (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bass</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tomlin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dulny</surname>
            <given-names>III</given-names>
          </string-name>
          , J.:
          <article-title>Quantum Annealing for Combinatorial Clustering</article-title>
          .
          <source>Quantum Information Processing</source>
          <volume>17</volume>
          (
          <issue>2</issue>
          ) (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>