<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Subjectively Interesting Alternative Clusters</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Tijl</forename><surname>De Bie</surname></persName>
							<email>tijl.debie@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Intelligent Systems Laboratory</orgName>
								<orgName type="institution">University of Bristol</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Subjectively Interesting Alternative Clusters</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B16415D0A7D8BA47E6D9251DE521EFA3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T13:42+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>Subjective interestingness, alternative clustering</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We deploy a recently proposed framework for mining subjectively interesting patterns from data <ref type="bibr" target="#b2">[3]</ref> to the problem of clustering, where patterns are clusters in the data. This framework outlines how subjective interestingness of patterns (here, clusters) can be quantified using sound information theoretic concepts. We demonstrate how it motivates a new objective function quantifying the interestingness of (a set of) clusters, automatically accounting for a user's prior beliefs and for redundancies between the clusters. Directly searching for the optimal set of clusters defined in this way is hard. However, the optimization problem can be solved to a provably good approximation if clusters are generated iteratively, paralleling the iterative data mining setting discussed in <ref type="bibr" target="#b2">[3]</ref>. In this iterative scheme, each subsequent cluster is maximally interesting given the previously generated ones, automatically trading off interestingness with non-redundancy. Thus, this implementation of the clustering approach can be regarded as a method for alternative clustering. Although generating each cluster in an iterative fashion is computationally hard as well, we develop an approximation technique similar to spectral clustering algorithms. We end with a few visual demonstrations of the alternative clustering approach to artificial datasets.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>A main challenge in research on clustering methods and theory is that clustering is (in a way intentionally) ill-defined as a task. As a result, numerous types of syntaxes for cluster patterns have been suggested (e.g. clusters as hyperrectangles, hyperspheres, ellipsoids, decision trees; clusterings as partitions, hierarchical partitionings, etc). Additionally, even when the syntax is fixed, there are often various alternative choices for the objective function (e.g. the K-means cost function, the likelihood of a mixture of Gaussians, etc).</p><p>Despite this variety in approaches, the goal of clustering is almost always to provide a user with insight in the structure of the data, allowing the user to conceptualize it as coming from a number of broad areas in the data space. The knowledge of such a structure can be more or less elucidating to the user, also depending on the prior beliefs the user held about the data.</p><p>Here, we take the perspective that a clustering is more useful if it conveys more novel information. We make a specific choice for a cluster syntax, and we deploy ideas from <ref type="bibr" target="#b2">[3]</ref> to quantify the interestingness of a cluster as the amount of information conveyed to the user when told about the cluster's presence.</p><p>Our approach attempts to quantify subjective interestingness of clusters, in that it takes prior beliefs held by the user into account. As a result, different clusters might be deemed interesting to different users. One particular example is the situation where a user has already been informed about previously discovered clusters in the data, which is the alternative clustering setting. In that case, clusters that are individually informative while non-redundant will be the most interesting ones. Our approach naturally deals with alternative clustering, by regarding already communicated clusters as prior beliefs.</p><p>Throughout this paper x ∈ R d denotes a d-dimensional data point, and X = x 1 x 2 • • • x n denotes the data matrix containing n data points as its rows. The space the data matrix belongs to is denoted as X = R n×d .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">A framework for data mining: a brief overview</head><p>For completeness, we here provide a short overview of a framework for data mining that was introduced in <ref type="bibr" target="#b2">[3]</ref>, and readers familiar with this paper can skip this section. Earlier and more limited versions of this framework, as well as its application to frequent pattern mining, can be found in <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b4">5]</ref>. For concreteness, here we specialize the short overview of the framework to the case where the data is a data set, summarized in the data matrix X.</p><p>The framework aims to formalize data mining as a process of information exchange between the data and the data miner (the user). The goal of the data miner is to get as good an understanding about the data as possible, i.e. to reduce his uncertainty as much as possible. To be able to do this, the degree of uncertainty must be quantified, and to this end we use probability distribution P (referred to as the background distribution) to model the prior beliefs of the user about the data X, in combination with ideas from information theory.</p><p>More specifically, the framework deals with the setting where the prior beliefs specify that the background distribution belongs to one of a set P of possible distributions. The more prior beliefs, the smaller this set will be. For example, the data miner may have a set of prior beliefs that can be formalized in the form of constraints the background distribution P satisfies:</p><formula xml:id="formula_0">X∈R n×d f i (X)P (X) = c i , ∀i.</formula><p>Such constraints represent the fact that the expected value of certain statistics (the functions f i ) are equal to a given number. The set P is defined as the set of distributions satisfying these constraints. (Note that the framework is not limited to such prior beliefs, although they are convenient from a practical viewpoint.)</p><p>We argued in <ref type="bibr" target="#b2">[3]</ref> that among all distributions P ∈ P, the 'best' choice for P is the one of maximum entropy given these constraints. This background distribution is the least biased one, thus not introducing any other undue constraints on the background distribution. A further game-theoretic argument in favour of using the distribution of maximum entropy is given in <ref type="bibr" target="#b2">[3]</ref>.</p><p>In the framework, a pattern is defined as any piece of knowledge about the data that reduces the set of possible values it may take from the original data space X = R n×d to a subset X . We then argued that the subjective interestingness of such a pattern can be adequately formalized as the self-information of the pattern, i.e. the negative logarithm of the probability that the pattern is present in the data, i.e. by − log (P (X ∈ X )). Thus, patterns are deemed more interesting if their probability is smaller under the background model, and thus if the user is more surprised by their observation.</p><p>After observing a pattern, a user instinctively adapts his beliefs. In <ref type="bibr" target="#b2">[3]</ref> we argued that a natural and robust way to model this is by updating the background distribution to a new distribution P defined as P conditioned on the pattern's presence. The self-information of subsequent patterns can thus be evaluated by referring to the new background distribution P , and so on in an iterative fashion.</p><p>In <ref type="bibr" target="#b2">[3]</ref> we showed that mining the most informative set of patterns formally corresponds to a weighted set coverage problem, attempting to cover as many elements from the set X that have a small probability under the initial background distribution P . This problem is NP-hard, but it can be approximated well by a greedy approach, and the iterative data mining approach is equivalent to such a greedy approximation.</p><p>Thus, the alternative clustering method we will detail below generates clusters in an iterative manner, in such a way that at any time the clusters generated so far are approximately the most informative set of clusters of that size.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Subjective interestingness of clusters</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Prior beliefs and the maximum entropy background distribution</head><p>Here we consider two types of initial prior beliefs, expressed as constraints on the first and second order cumulants of the data points. It is conceptually easy to extend the results from this paper to other types of prior beliefs, although the computational cost will vary. The background distribution incorporating these constraints is the maximum entropy distribution that has the prescribed first and second order cumulants. It is well known that for data with unbounded domain, this distribution is the multivariate Gaussian distribution with mean and covariance matrix equal to the prescribed cumulants:</p><formula xml:id="formula_1">P (X) = 1 (2π) nd |Σ| n exp − 1 2 trace (X − eμ )Σ −1 (X − eμ ) . (<label>1</label></formula><formula xml:id="formula_2">)</formula><p>We note that the prescribed cumulants may be computed from the actual data at the request of the data miner, such that they are indeed part of the prior knowledge. However, they may also be beliefs, in the sense that they may be based on external information or assumptions that may be right or wrong.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">A syntax for cluster patterns</head><p>The framework from <ref type="bibr" target="#b2">[3]</ref> was developed for patterns generally defined as properties of the data. Thus, a pattern's presence in the data constrains the set of possible values the data may have, and in this sense the knowledge of the presence of a pattern reduces the uncertainty about the data and conveys information.</p><p>In this paper we restrict our attention to one specific type of cluster pattern. The pattern type we consider is parameterized by a set of indices to the data points and a vector in the data space. The pattern is then the fact that the mean of the data points for these indices is equal to the specified vector.</p><p>More formally, let e I ∈ R d be defined as an indicator vector containing zeros at positions i ∈ I and ones at positions i ∈ I, and let n I = |I| = e I e I denote the number of elements in I. Then, our patterns are constraints of the form:</p><formula xml:id="formula_3">1 n I i∈I x i = μ I , ⇔ X e I e I e I = μ I .</formula><p>Such a constraint restricts the possible values of the data set X, in that the mean of a subset of the data points is required to have a prescribed value μ I .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">The self-information of a cluster pattern</head><p>The following theorem shows how to assess the self-information of a pattern.</p><p>Theorem 1. Given a background distribution of the form in Eq. ( <ref type="formula" target="#formula_1">1</ref>), the probability of a pattern of the form X eI e I eI = μ I is given by:</p><formula xml:id="formula_4">P X e I e I e I = μ I = 1 (2π) d |Σ| exp − 1 2|I| e I • (X − eμ )Σ −1 (X − eμ ) • e I .</formula><p>Thus the self-information of the pattern for a cluster specified by the set I, defined as the negative log probability of the cluster pattern and denoted as SelfInformation I , is equal to:</p><formula xml:id="formula_5">SelfInformation I = 1 2 log (2π) d |Σ| + 1 2 Q I ,</formula><p>where</p><formula xml:id="formula_6">Q I = 1 |I| e I • (X − eμ )Σ −1 (X − eμ ) • e I .</formula><p>Note that the self-information depends on I only through Q I , so we may choose to use Q I as a quality metric for a cluster, as we will do in this paper. This theorem can be used to quantify the self-information of any cluster given the background model based on the prior beliefs of the data miner. Note however that it cannot be used to assess the self-information of a cluster after other clusters have already been found, as each new cluster will affect the user's prior beliefs. How this can be accounted for will be discussed in Sec. 3.4, based on a generalization of Theorem 1. As Theorem 1 directly follows from Theorem 2, we will only provide a proof for the latter in Sec. 3.4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">The self-information of a set of cluster patterns</head><p>Let us discuss the more general case, where the patterns are constraints of the form X E = M with E ∈ R n×k and M ∈ R d×k . Clearly, the type of patterns we wish to consider are a special case of this, with E = eI e I eI and M = μ I . Furthermore, it allows us to consider a composite pattern, a pattern defined as the union of a set of k patterns. Indeed, if we have k different cluster patterns specified by the sets from I = {I i }, we can write down this set of constraints concisely as X E = M where E and M contain eI i e I i eI i and the mean vector μ i of the i'th cluster as their i'th columns.</p><p>Theorem 2. Let the columns of the matrix E be the indicator vectors of the sets in I = {I i }, and let P E = E(E E) −1 E , the projection matrix onto the column space of E. Then, the probability of the composite pattern X E = M is given by:</p><formula xml:id="formula_7">P (X E = M) = 1 (2π) kd |Σ| k exp − 1 2 trace P E • (X − eμ )Σ −1 (X − eμ ) .</formula><p>Thus the self-information of the set of patterns defined by the columns of E, defined as its negative log probability and denoted as SelfInformation I , is equal to:</p><formula xml:id="formula_8">SelfInformation I = k 2 log (2π) d |Σ| + 1 2 Q I ,</formula><p>where</p><formula xml:id="formula_9">Q I = trace P E • (X − eμ )Σ −1 (X − eμ ) .</formula><p>Again, since the self-information depends on I only through Q I , we choose to use Q I as a quality metric for a cluster further below.</p><p>Proof. A constraint X E = M constrains the data X to an (n − k) × d dimensional affine subspace in the following way. Let us write the singular value decomposition for E as:</p><formula xml:id="formula_10">E = U U 0 Λ 0 0 0 V V 0 .</formula><p>Then, this constraint can be written in the following form:</p><formula xml:id="formula_11">X = UZ + U 0 Z 0 ,</formula><p>where Z = Λ −1 V M is a constant fixed by E and M, and Z 0 ∈ R (n−k)×d is a variable. In general, writing X = UZ + U 0 Z 0 , we can write the probability density for X as:</p><formula xml:id="formula_12">P (X) = P (Z, Z 0 ), = 1 (2π) nd |Σ| n exp − 1 2 trace (UZ + U 0 Z 0 − eμ )Σ −1 (UZ + U 0 Z 0 − eμ ) , = 1 (2π) (n−k)d |Σ| n−k exp − 1 2 trace (Z 0 − U 0 eμ )Σ −1 (Z 0 − U 0 eμ ) • 1 (2π) (k)d |Σ| k exp − 1 2 trace (Z − U eμ )Σ −1 (Z − U eμ )</formula><p>We can now compute the marginal probability density for Z by integrating over Z 0 , yielding:</p><formula xml:id="formula_13">P (Z) = 1 (2π) kd |Σ| k exp − 1 2 trace (Z − U eμ )Σ −1 (Z − U eμ ) .</formula><p>The probability density value for the pattern's presence, i.e. for X E = M or equivalently Z = Λ −1 V M , is thus:</p><formula xml:id="formula_14">P (Z = Λ −1 M ) = 1 (2π) kd |Σ| k exp − 1 2 trace (Λ −1 V M − U eμ )Σ −1 (Λ −1 V M − U eμ ) , = 1 (2π) kd |Σ| k exp − 1 2 trace P E • (X − eμ )Σ −1 (X − eμ ) ,</formula><p>where</p><formula xml:id="formula_15">P E = E(E E) −1 E is a projection matrix projecting onto the k-dimensio- nal column space of E.</formula><p>Note that Theorem 1 is indeed a special case of Theorem 2 as can be seen by substituting E = e I and P E = eI e I |I| . According to the framework, a (composite) pattern specified by matrices E and M in this way is thus more informative if trace P E • (X − eμ )Σ −1 (X − eμ ) is larger. Thus, we could search for the most informative set of clusters by maximizing this quality measure with respect to a set I = {I i } of clusters. This is a hard problem though, even in the case where only one cluster is sought. Thus, we developed an approximation algorithm.</p><p>Our approach is approximate in two ways. First, the search for a set of clusters is approximated using a greedy algorithm, searching clusters one by one, thus operating like an alternative clustering algoritm. From <ref type="bibr" target="#b2">[3]</ref>, it can be seen that this greedy approximation is guaranteed to approximate the true optimum provably well. Second, the search for each cluster is relaxed to an eigenvalue problem. These issues are discussed in greater detail in Sec. 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5">The cost of describing a cluster</head><p>The framework from <ref type="bibr" target="#b2">[3]</ref> suggests to take into account not only the self-information of a pattern, but also the cost to communicate a pattern, i.e. its description length. This depends on the coding scheme used, which should reflect the perceived complexity of a pattern as perceived by the data miner. Choosing this coding scheme can also be done so as to bias the results toward specific patterns.</p><p>In the current context, describing a pattern amounts to describing the subset I and the mean vector μ I . For simplicity, we assume the description length is constant for all patterns, independent of I and μ I . However, note that different costs could be used if patterns with smaller sets I are more easy to understand (i.e. have a smaller cost), or vice versa.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Alternative clustering: finding the next most informative cluster</head><p>Here we discuss an iterative approach to optimizing the quality measure Q I from Theorem 2. There are two reasons for choosing an iterative approach. Firstly, directly optimizing the quality measure is equivalent to a set covering type problem (see <ref type="bibr" target="#b2">[3]</ref> for more background on why this is the case). While NPhard, this problem can be approximated well by optimizing over the different clusters (and thus the columns of E) in a greedy iterative manner.</p><p>Secondly, usually it is not a priori clear how many clusters are required for the data miner to be sufficiently satisfied with his new understanding of the data. The idea of alternative clustering, as we view it, is to provide the user the opportunity to request new clusters (or clusterings) as long as more are desired. Optimizing the quality measure over a growing set of clusters by iteratively optimizing over a newly added column of E is thus a type of alternative clustering.</p><p>Hence, the iterative approach can be regarded as an approximation, but one with usability benefits over a global optimizing approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">The iterative scheme: alternative clustering</head><p>To search for the first cluster, we attempt to optimize the quality function from Theorem 1. This is itself a hard problem, but we explain how we approximately solve it in Sec. 4.2.</p><p>In the subsequent iterations, let us say that we have already found k − 1 ≥ 1 clusters, and the matrices E and M respectively contain the normalized indicator vectors and cluster means as their columns. We are interested in finding the k'th cluster so as to optimize the quality measure from Theorem 2 but keeping the first k − 1 cluster patterns as they are.</p><p>To do this, it is convenient to write the quality measure as a function of the k'th cluster with indicator vector e k . Let Q E = I − P E , the projection matrix on the null column space of E. Furthermore, let us denote E * = E e k . Then, using the definition of a projection matrix and the matrix inversion lemma:</p><formula xml:id="formula_16">Q {Ii|i=1:k} = trace P E * • (X − eμ )Σ −1 (X − eμ ) , = trace P E • (X − eμ )Σ −1 (X − eμ ) +trace Q E e k e k Q E e k Q E e k • (X − eμ )Σ −1 (X − eμ ) , = Q {Ii|i=1:k−1} + e k Q E • (X − eμ )Σ −1 (X − eμ ) Q E e k e k Q E e k .</formula><p>Note that if we define Q ∅ = 0 and with Q E = I the quality measure from Theorem 1 is retrieved. We can thus interpret the above reformulation of the quality metric for the k'th cluster conditioned on the first k − 1 clusters as being the quality metric for a first cluster on data that is projected onto the space orthogonal to the k − 1 columns of E, i.e. the k − 1 previously selected indicator vectors. It is as if the data was deflated to take account of the knowledge of the previously found cluster patterns, thus automatically accounting for redundancy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">A spectral relaxation of the iterations</head><p>Each of the iterative steps thus reduces to the maximization of the following increase of the quality measure:</p><formula xml:id="formula_17">ΔQ k = e k Q E • (X − eμ )Σ −1 (X − eμ ) Q E e k e k Q E e k . (<label>2</label></formula><formula xml:id="formula_18">)</formula><p>If we relax the vector e k to be real-valued instead of containing only 0's and 1's, this Rayleigh quotient is maximized by the dominant eigenvector of the matrix</p><formula xml:id="formula_19">Q E • (X − eμ )Σ −1 (X − eμ ) Q E .</formula><p>Thus, as an approximation technique we will use this dominant eigenvector, and threshold it to obtain a crisp 0/1 vector e k . To determine a suitable threshold, we simply do an exhaustive search over n + 1 threshold values that generate a different set I of indices i ∈ I for which e k (i) = 1, selecting the threshold that maximizes the quantity in Eq. ( <ref type="formula" target="#formula_17">2</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">A kernel-based version</head><p>Note that for Σ = I, the quality metrics depend on X only through the inner product matrix XX . This means that a kernel-variant is readily derived, by substituting this inner product matrix with any suitable kernel matrix. In this way nonlinearly shaped clusters can be obtained, similar to spectral clustering methods and kernel K-Means.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Relations to existing work</head><p>There appear to be strong relations between spectral clustering and our spectral relaxation of the problem <ref type="bibr" target="#b6">[7]</ref>. Additionally, the quality measure is strongly related to the K-Means cost function <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b5">6]</ref>. Finally, there seem to be interesting connections to (0-1) SDP problems used for solving combinatorial optimization problems such as clustering (e.g. K-Means and graph cut clustering) <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b0">1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Experiments</head><p>We conducted 3 experiments, each time reporting the result of 6 iterations of the alternative clustering scheme. Initial prior beliefs are always μ = 0 and Σ = I.</p><p>-A plain application to a synthetic dataset of 100 points and 2 dimensions in 4 clusters. See Fig. <ref type="figure">1</ref>. -An application to the same data but now with a Radial Basis Function (RBF) kernel used for the inner products. See Fig. <ref type="figure" target="#fig_1">2</ref>. -An application with an RBF kernel to a different synthetic dataset, with one central cluster and two half-moon shaped clusters around this central cluster. See Fig. <ref type="figure">3</ref>. A synthetic dataset with 25 data points sampled from each of 4 2-dimensional Gaussian distributions with identity covariance matrix and different means. From left to right and top to bottom, the plots show the first 6 consecutive alternative clusters found by our method when a standard inner product is used (data points belonging to the cluster are plotted using crosses). The numbers above the plots are the values of ΔQ k from Eq. (2) for the cluster shown. Note that it is high for the first 3 clusters, which reveal the enforced cluster structure, before dropping to a much lower level. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions</head><p>In <ref type="bibr" target="#b2">[3]</ref> we introduced a framework for data mining, aiming to quantify the subjective interestingness of patterns. We showed that Principal Component Analysis can be seen as implementing this framework for a particular pattern type and prior beliefs, thus providing an alternative justification for this method. In earlier work we also showed the potential of the framework in quantifying subjective interestingness for frequent itemset mining <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5]</ref>. Now, in the present paper, we showed in detail how the framework can also be applied successfully to the case of clustering, leading to a new approach for alternative clustering that presents subjectively interesting clusters in data in an iterative data mining scheme.</p><p>In further work, we will investigate the quality of the spectral relaxation, and consider the development of tighter relaxations (e.g. to semi-definite programs). We will also further develop links with spectral clustering and other existing clustering approaches, to provide alternative justifications and insights or to improve on these approaches. We will also investigate the use of other pattern syntaxes for cluster(ing)s, and the use of more complex types of prior beliefs. Lastly, we plan to demonstrate the power of the framework by also applying these ideas to other types of data and corresponding types of prior beliefs, such as positive real-valued, integer, binary, and more structured types of data.</p><p>Due to space constraints, in this workshop paper we could not situate the contributions within the wider literature on alternative clustering. We will rectify this important shortcoming in a later version of this workshop paper.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Fig.1. A synthetic dataset with 25 data points sampled from each of 4 2-dimensional Gaussian distributions with identity covariance matrix and different means. From left to right and top to bottom, the plots show the first 6 consecutive alternative clusters found by our method when a standard inner product is used (data points belonging to the cluster are plotted using crosses). The numbers above the plots are the values of ΔQ k from Eq. (2) for the cluster shown. Note that it is high for the first 3 clusters, which reveal the enforced cluster structure, before dropping to a much lower level.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. A synthetic dataset with 25 data points sampled from each of 4 2-dimensional Gaussian distributions with identity covariance matrix and different means. From left to right and top to bottom, the plots show the first 6 consecutive alternative clusters when an RBF kernel with kernel width 3 is used. The numbers above the plots are the values of ΔQ k from Eq. (2) for the cluster shown.</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work is supported by the EPSRC grant EP/G056447/1.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"> <ref type="formula">2</ref><p>) for the cluster shown. Note that the second cluster generated contains all data points. This is possible and sensible from the perspective of our approach if the mean of the entire data set is significantly different from the expected mean in the initial background model. This may well be the case when working in a Hilbert space induced by the RBF kernel, where all data points lie in one orthant such that their mean cannot be in the origin.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Fast sdp relaxations of graph cut clustering, transduction, and other combinatorial problems</title>
		<author>
			<persName><forename type="first">T</forename><surname>De Bie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Cristianini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="1409" to="1436" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Maximum entropy models and subjective interestingness: an application to tiles in binary databases</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">De</forename><surname>Bie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Mining and Knowledge Discovery</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">An information-theoretic framework for data mining</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">De</forename><surname>Bie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD11)</title>
				<meeting>of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD11)</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A framework for mining interesting pattern sets</title>
		<author>
			<persName><forename type="first">T</forename><surname>De Bie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-N</forename><surname>Kontonasios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Spyropoulou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGKDD Explorations</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">An information-theoretic approach to finding informative noisy tiles in binary databases</title>
		<author>
			<persName><forename type="first">K.-N</forename><surname>Kontonasios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">De</forename><surname>Bie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2010 SIAM International Conference on Data Mining</title>
				<meeting>the 2010 SIAM International Conference on Data Mining</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Approximating k-means-type clustering via semidefinite programming</title>
		<author>
			<persName><forename type="first">Jiming</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Wei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIAM Journal on Optimization</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Normalized cuts and image segmentation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Malik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="888" to="905" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Spectral relaxation for k-means clustering</title>
		<author>
			<persName><forename type="first">H</forename><surname>Zha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Simon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2002">NIPS01. 2002</date>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="1057" to="1064" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
