<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Factorial Clustering with an Application to Plant Distribution Data</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Manfred</forename><surname>Jaeger</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science</orgName>
								<orgName type="institution">Aalborg University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Simon</forename><forename type="middle">P</forename><surname>Lyager</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science</orgName>
								<orgName type="institution">Aalborg University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Michael</forename><forename type="middle">W</forename><surname>Vandborg</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Computer Science</orgName>
								<orgName type="institution">Aalborg University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Thomas</forename><surname>Wohlgemuth</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Swiss Federal Research Institute WSL</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Factorial Clustering with an Application to Plant Distribution Data</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">EB69979BCA0E8630CCC967946622ACC8</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T13:42+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We propose a latent variable approach for multiple clustering of categorical data. We use logistic regression models for the conditional distribution of observable features given the latent cluster variables. This model supports an interpretation of the different clusterings as representing distinct, independent factors that determine the distribution of the observed features. We apply the model for the analysis of plant distribution data, where multiple clusterings are of interest to determine the major underlying factors that determine the vegetation in a geographical region.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>There exist a variety of different approaches to learning multiple clusterings. They can differ not only with regard to their mathematical models and algorithmic methods, but there can also be widely different intuitions and objectives with regard to the interpretation of the multiple clusterings. On the one hand, in ensemble clustering, the individual clusterings are essentially regarded as different, imperfect versions of a single underlying true clustering (e.g. <ref type="bibr" target="#b9">[10]</ref>). In many multiple clustering methods, on the other hand, the different clusterings are intended to represent different views of the data, each providing a different insight into the structure of the data. One objective for clustering methods then is to ensure that different clusterings are in some sense independent, disparate <ref type="bibr" target="#b5">[6]</ref>, or non-redundant <ref type="bibr" target="#b7">[8]</ref>.</p><p>Probabilistic latent variable models are used for a variety of data analysis tasks (in the case of discrete data jointly referred to as latent class analysis), including clustering. Several authors have investigated probabilistic latent variable models for multiple clusterings <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b1">2]</ref>. For the case of discrete observable features, no special assumptions on the distributional form of the features given the latent variables are made in these approaches, i.e. the conditional distribution of the features follows an unconstrained multinomial distribution. Latent variable models are also commonly used for dimensionality reduction of high-dimensional numeric data. An important example is the factor analysis model, in which the observed data is interpreted as a noisy linear transformation of a small number of latent dimensions. While it is quite common to refer to latent class analysis as a "categorical data analogue to factor analysis" <ref type="bibr" target="#b4">[5]</ref>[1, <ref type="bibr">Chapter 13]</ref>, it seems that this correspondence has not been fully exploited for clustering applications, or put into the context of multi-clustering, via the combined use of multiple latent variables, and special assumptions on the conditional distribution of the observed features.</p><p>In this paper we propose a probabilistic latent variable model for multiple clusterings. As in factor analysis, we interpret the observed (discrete) data as a noisy transformation of underlying, discrete latent dimensions. The linear mapping of factor analysis is replaced by a log-linear logistic regression model. The latent dimensions then define clusterings that can be seen as independent factors that determine the distribution of the observed features.</p><p>In contrast with several other multiple clustering methods (e.g., <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b7">8]</ref>) our method is not based on an association of different clusterings with different feature subsets, even though such associations can emerge.</p><p>Our approach is partly motivated by applications to biogeographical data. Specifically, we are investigating plant distribution data. Segmentations of geographic units into floristic regions based on similarity of plant species composition were already undertaken in the 19th century. An early application of formal methods of clustering in this context is <ref type="bibr" target="#b8">[9]</ref>. We apply our method to distribution data for 2398 plant species in Switzerland. The goal of factorial clustering for this type of data will be to obtain multiple clusterings, each of which could correspond to one of several underlying environmental, geographical, or historical factors, which jointly influence the vegetation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Latent Variable Models for Clustering</head><p>Latent variable models are routinely used for clustering, both for single and multiple clustering. However, they can be used in several, slightly different ways. In order to more clearly explain our approach, we briefly review in this section possible approaches to using latent variable models for clusterings.</p><p>Throughout, we assume that the observable data X consists of n observations of k attributes, i.e. X is an n × k matrix. A latent variable model contains m additional unobserved variables, and we denote with L the n × m matrix of the latent variables in the n observations. We note that when we assume that in the n observations both the observable and latent variables are identically and independently sampled, it will be simpler and more natural to describe the model in terms of vectors X, L of length n and m, respectively. However, in some applications, especially segmentation of time sequences or images, the latent variables are not independent at different data points.</p><p>A latent variable model, then, consists of a joint distribution for X and L, which can be written as</p><formula xml:id="formula_0">P (X | L, θ X|L )P (L | θ L ).<label>(1)</label></formula><p>In hierarchical models, this might be extended by a distribution over θ X|L , θ L parametrized by hyperparameters λ.</p><p>The perhaps most common use of model <ref type="bibr" target="#b0">(1)</ref> for clustering is to perform two steps <ref type="bibr" target="#b12">[13]</ref>: first, fit the parameters θ X|L , θ L by maximizing the marginal likelihood of the observed data X = x:</p><formula xml:id="formula_1">(θ * X|L , θ * L ) := arg max θ X|L ,θL l P (X = x | L = l, θ X|L )P (L = l | θ L ). (<label>2</label></formula><formula xml:id="formula_2">)</formula><p>This step is usually performed using the EM algorithm. Then, compute the most probable values of L given X = x:</p><formula xml:id="formula_3">l * = arg max l P (L = l | X = x, θ * X|L , θ * L )<label>( 3 )</label></formula><p>In multiple clustering, a joint configuration of the latent variables defines multiple cluster indices. For simplicity we may assume for now that each latent variable defines its own clustering, and that therefore the membership of the ith data item in the jth clustering is given by l * i,j . However, in the multi-cluster case, the second step can also take a slightly different form, and the most probable latent variable values be computed component-wise. Denoting by l j the jth column of l (i.e., l = (l 1 , . . . , l m )), this can be written as l * j = arg max lj l1,...,lj−1,lj+1,lm</p><formula xml:id="formula_4">P (L = l | X = x, θ * X|L , θ * L ).<label>(4)</label></formula><p>This is the (hard) clustering rule used, e.g., in <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14]</ref>. The clusterings obtained from ( <ref type="formula" target="#formula_3">3</ref>) and (4) can differ.</p><p>If the ultimate goal is only to compute a most probable configuration of L, then one may also try to simplify the combination of (2) and (3) into a single optimization:</p><formula xml:id="formula_5">l * := arg max l max θ X|L ,θL P (X = x | L = l, θ X|L )P (L = l | θ L ).</formula><p>(</p><formula xml:id="formula_6">)<label>5</label></formula><p>This rule can be justified by a Bayesian interpretation, for example: it amounts to finding the jointly most probably values of l, θ X|L , θ L , given the data X = x, and assuming a uniform prior for θ X|L , θ L . Rule (5) may be still further simplified, if one assumes the model for the latent variables to be fixed, and not subject to optimization, i.e., P (L = l | θ L ) = P (L = l | θ * L ) for fixed parameters θ * L , and the parameter optimization is only for θ X|L . If, furthermore, P (L = l | θ * L ) is assumed uniform, then (5) reduces to</p><formula xml:id="formula_7">l * := arg max l max θ X|L P (X = x | L = l, θ X|L ). (<label>6</label></formula><formula xml:id="formula_8">)</formula><p>Whether it is justified to assume a fixed distribution P (L | θ * L ) can depend on two considerations: first, assuming that (1) actually represents the generative process for the data, one might have sufficient background knowledge to identify the distribution of L a-priori. L being an unobserved variable, whose existence is essentially hypothesized, and for which it is typically even unclear how many states it has, this is a rather unlikely case in practice, however. Second, clustering being an exploratory data-analysis tool, one may also consider what settings of P (L | θ * L ) may lead via <ref type="bibr" target="#b4">(5)</ref> to interesting insights into the data, regardless of whether the underlying probabilistic model is accurate as a generative model.</p><p>For example, in the single clustering case, when the data is generated by a mixture model where one mixture component has a much higher prior probability than the others, then clustering via (3) can easily lead to only obtaining a single cluster. If, on the other hand, one eliminates the influence of the prior distribution by assuming (incorrectly) a uniform distribution over the mixture components, then clustering via (6) can reveal the mixture structure of the data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">The Factorial Logistic Model</head><p>In the factor analysis model, both X and L are numerical, the rows in X and L are iid, and the model ( <ref type="formula" target="#formula_0">1</ref>) is given by distribution</p><formula xml:id="formula_9">P (L i ) ∼ N (0, Σ L ) P (X i | L i ) ∼ N (W L i + μ, Σ X )</formula><p>where Σ L is an arbitrary covariance matrix, W is a k × m matrix, μ a mdimensional mean vector, and Σ X a diagonal covariance matrix. Thus, data is assumed to be generated by sampling from a lower (k) dimensional Gaussian distribution, linearly mapped into the higher (m) dimensional space, and independent Gaussian noise added to each coordinate.</p><p>The logistic regression model for the distribution of a binary variable X conditional on numeric latent variables L is given by log</p><formula xml:id="formula_10">P (X = 1 | L)/P (X = 0 | L) = w 0 + wL,<label>(7)</label></formula><p>where w = (w 1 , . . . , w k ) is a k-vector of real weights. We write X ∼ LR(w 0 , wL) if X follows <ref type="bibr" target="#b6">(7)</ref>. This model also applies when the latent variables L are ordinal, i.e. each L j codes by an integer {0, . . . , r j − 1} one of r j different, ordered categories. To accommodate nominal predictor variables (i.e., unordered categorical variables) in the logistic regression model, one encodes a nominal variable L j with r states by binary indicator variables L j,1 , . . . , L j,r , i.e. L j,h = 1, L j,h = 0 (h = h) means that L j is in its hth state. We will consider both ordinal and nominal latent variables for clustering. An ordinal latent variable defines an ordered clustering, i.e. the cluster indices define an ordering of the clusters. Whether such an ordering is meaningful and interpretable is application dependent. For biogeographical data ordinal latent variables and ordered clusterings are often natural, since data patterns are often determined by underlying continuous variables. We will, thus, assume that L is a vector of m latent variables that define c different clusterings. Furthermore, we assume that one of the following two cases applies: (1) all L j in L are ordinal; in this case c = m, and the jth clustering consists of r j distinct cluster. (2) L is an encoding by binary indicator variables of c distinct nominal variables with r 1 , . . . , r c distinct states, respectively. In this case m = c i=1 r i . We refer to model (1) as the (r 1 o, . . . , r k o) model, and (2) as the (r 1 n, . . . , r c n) model. One could also consider models combining ordinal and nominal latent variables, but we will here focus on "pure" models.</p><p>As in the factor analysis model, we assume that</p><formula xml:id="formula_11">P (X | L) ∼ n i=1 P (X i | L i ) ∼ n i=1 k j=1 P (X i,j | L i ).</formula><p>Assuming that each X i,j follows a logistic regression model <ref type="bibr" target="#b6">(7)</ref> with parameters w j,0 , w j , one obtains the model for the ith data item:</p><formula xml:id="formula_12">P (X i | L i ) ∼ k j=1 LR(w j,0 , w j L i ).<label>(8)</label></formula><p>This conditional model for X may be combined with various models for P (L), with or without an iid assumption for the rows of L. We refer to multiple clustering based on (8) as factorial logistic (FL) clustering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Learning</head><p>We apply the simple learning rule (6) for clustering with the logistic regression model. Thus, we assume that L is uniformly distributed, which implies, in particular, independence over rows: P (L) ∼ i P (L i ). In case of L encoding nominal variables, the uniform distribution, of course, is conditional on "legal" states of L, i.e. at most one indicator variable for any particular nominal variable being equal to 1.</p><p>For the optimization of ( <ref type="formula" target="#formula_7">6</ref>) we then use the obvious iterative procedure, where after a random initialization L := l 0 two steps are alternated:</p><formula xml:id="formula_13">i θ X|L t := arg max θ X|L P (X = x | L = l t , θ X|L ) ii l t+1 := arg max l P (X = x | L = l, θ X|L t )</formula><p>Step i is performed in our implementation using the SPSS method of fitting logistic regression models, which supports both ordinal and nominal predictor variables. Due to the factorization (8), the optimization reduces to k independent optimizations for the parameters (w j,0 , w j ) (j = 1, . . . , k). It is thus linear in k. It also is linear in n, since the likelihood only depends on the counts |{i | X i,j = 1, L i = l} | for fixed configurations l of the latent variables.</p><p>For step ii we have</p><formula xml:id="formula_14">P (X = x | L = l, θ X|L t ) = i P (X i = x i | L i = l i , θ X|L t )</formula><p>, so that the problem decomposes into n distinct optimizations for the l i . It can be naively performed by computing P (X i = x i | L i = l i , θ X|L t ) for each candidate l i , which gives a procedure that is still linear in n and k, but exponential in c.</p><p>Overall, we obtain a learning method that is linear in the number of data items and the observable attributes, and exponential in the number of clusterings. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments</head><p>We apply FL-clustering to geobotanical data. In our experiments we use the source data for the "Swiss Web Flora"<ref type="foot" target="#foot_0">3</ref>  <ref type="bibr" target="#b10">[11]</ref>. The dataset contains information on the distribution of 2697 plant species in Switzerland, which has been divided into 565 mapping areas. We reduced slightly more detailed species abundance information in the original data to simple binary presence/absence data. We also in this process deleted plants with a very sparse and uncertain distribution. This left us with 2398 species in our data. <ref type="foot" target="#foot_1">4</ref> We view each plant species as an observable attribute, and the mapping areas as independent observations. Thus, n = 565 and k = 2398 in the notation of Section 2. Figure <ref type="figure" target="#fig_0">1</ref> shows the division of Switzerland into the mapping areas. Apart from the species occurrence data, only a single additional variable is recorded for each area: a binary variable that indicates whether the area is a mountain area (above timberline), or a valley area (below timberline). The value of this variable is shown if Figure <ref type="figure" target="#fig_0">1</ref> by a green color for valley, and grey color for mountain areas.</p><p>Conventional (single-) clusterings of the data lead to a segmentation of Switzerland into floristic regions. Figure <ref type="figure" target="#fig_0">1</ref> (b) shows a result obtained by agglomerative hierarchical clustering of the valley areas only <ref type="bibr" target="#b11">[12]</ref> (thus, the white part of the figure does not correspond to a computed cluster; it comprises areas not included in the clustering).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Synthetic Data</head><p>In order to obtain an initial evaluation of the feasibility of our approach, we first conduct an experiment with synthetic data. For this we constructed two artificial segmentations of Switzerland based on the same mapping regions as in the real data. These segmentations are shown in Figure <ref type="figure" target="#fig_1">2</ref> (top), and henceforth referred to as "vertical" and "horizontal" segmentation, respectively. For each combination of a vertical and a horizontal segment, we defined a species distribution type by a nominal logistic regression model that expresses a preference of the species for the selected vertical and horizontal segment. The logistic regression weights were adjusted so as to obtain conditional probability distributions for the presence of a species of the following form (here showing the case of preference for the first segment in both segmentations): Vertical Horizontal yellow green blue blue 0.98 0.5 0.5 yellow 0.5 0.02 0.02 <ref type="bibr" target="#b8">(9)</ref> According to each distribution type we created 15 synthetic plant species, and randomly sampled an occurrence variable for the species at each of the mapping areas.</p><p>We then performed FL clustering based on the 90 synthetic species using the (3o,2o) model (it is not our ambition at this point to detect the "right" number of segmentations and segments per segmentation). In approximately 1 out of 3 random restarts the algorithm terminated with the correct segmentations of Figure <ref type="figure" target="#fig_1">2</ref>, or solutions that differed from the correct one in cluster assignments for 2-3 regions. In the remaining restarts the algorithm terminated at local optima, a representative example of which is shown in Figure <ref type="figure" target="#fig_1">2</ref> (bottom). However, the (almost) correct solutions were identified by higher log-likelihood scores <ref type="bibr">(between -19912 and -19783)</ref> than that of the wrong solutions (between -23474 and -21677).</p><p>For comparison, we also performed an experiment where the logistic regression model for P (X | L) was replaced by a full multinomial model, i.e. for each species we fit a conditional probability table of the form (9) with 6 independent parameters. In this case, almost all restarts terminated with wrong solutions as in Figure <ref type="figure" target="#fig_1">2</ref>, and, more importantly, the correct solutions could not be distinguished by a higher likelihood score: in the multinomial model, any pair of segmentations whose combination identifies the 6 different combinations of vertical and horizontal segments achieves the same, optimal, likelihood score.</p><p>We also use this synthetic data experiment to demonstrate that in FLclustering there is not necessarily a correlation between clusterings and featuresubsets. Figure <ref type="figure">3 (a)</ref> shows for each of the 90 synthetic species the mutual information between the species occurrence feature and the two clusterings of Figure <ref type="figure" target="#fig_1">2</ref> (top). The plot shows that there is no strong association of individual species features with one or the other of the two segmentations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Real Data</head><p>We now perform experiments with the real data consisting of the actual 2398 species. Again, we do not try at this point to automatically detect an appropriate number of segmentations, or segments per segmentation. We run the learning algorithm with a few selected ordinal and nominal logistic models. In all cases we perform 20 runs of the algorithm with different random initializations of the latent variables L. The results shown in the following are the segmentations that achieved the highest likelihood score <ref type="bibr" target="#b5">(6)</ref> within the 20 restarts.</p><p>Figure <ref type="figure">4</ref> shows the result of clustering with the (3o,3o) and (3n,3n) logistic models. We use different colors to represents segments computed by nominal logistic models, and greyscale values for ordinal logistic models. The greyscale values then show the ordering of the segments according to their (ordinal) index. One first observes that both models have produced one segmentation in which the mountain areas are identified as one segment: there is an almost perfect correspondence between the mountain attribute illustrated in Figure <ref type="figure" target="#fig_0">1</ref>, and the dark grey, respectively yellow, segments in the first segmentations of Figure <ref type="figure">4</ref>  (note that our method does not entail an ordering of the different segmentations; in particular, in Figure <ref type="figure">4</ref> we have just for convenience vertically aligned similar segmentations, and arbitrarily put the ones containing the mountain segment first).</p><p>Apart from the mountain/valley attribute our data does not contain "hidden class variables" that could be used for interpreting the segmentations, and therefore one has to look for additional, external data sources, and expert knowledge. As previously mentioned, we expect that the different segmentations to some extent correspond to ecological factors that determine plant growth. A difficulty we now encounter is that many such candidate factors (e.g., average annual temperature, average precipitation) are highly correlated with the mountain/valley division, and often show a secondary gradient in north-south direction. The second segmentations of Figure <ref type="figure">4</ref> are somewhat dominated by a north-south stratification, and also exhibit some of the patterns visible in Figure <ref type="figure" target="#fig_0">1 (b</ref>). However, it seems impossible to identify these north-south segmentations with any particular ecological factor. Instead, it can only be taken as aggregating the northsouth dependency of several factors. Moreover, whereas one clustering showing the mountain/valley division was quite consistently produced in the random restarts, there were larger variations observed in the north-south clustering.</p><p>The range of likelihood values obtained in 20 restarts was −288 • 10 3 to −278•10 3 for (3o,3o) clustering, and −308•10 3 to −296•10 3 for (3n,3n) clustering (since the former model fits more independent parameters, higher likelihood scores are to be expected).</p><p>Figure <ref type="figure">3</ref> shows the mutual information values for the plant features and the (3o,3o) segmentation of Figure <ref type="figure">4</ref>. This plot shows a relatively strong cor-relation of some species with the mountain/valley clustering, and a somewhat less pronounced primary correlation of some other species with the north/south segmentation. The large number of species with very small mutual information values for either segmentation is largely made up of species that occur in only a few areas. In analogy to the experiment with synthetic data, we also perform with the real data an experiment with full multinomial species distribution models instead of the logistic ones. The result is shown in Figure <ref type="figure" target="#fig_3">5</ref>. While the mountain/valley pattern is also partly visible in some of the segments, there is no single segment or segmentation corresponding to this division, and the overall segmentation result is clearly less useful than the one obtained with the logistic models.</p><p>The poor performance of the multinomial model may in part be due to this model's inability to isolate in its different clusterings several independent explanatory factors, as illustrated by the synthetic data experiments. In addition, the multinomial model suffered from severe problems of convergence to local optima: even though the global likelihood maximum of the multinomial model must be at least as high as the logistic optimum, the likelihood values found in 20 restarts were significantly lower for the multinomial than for the logistic models (range −433 • 10 3 to −405 • 10 3 ).</p><p>The average time consumption of a single run (restart) of (3o,3o) or (3n,3n) clustering was approximately 3 hours, with an average of approximately 15 iterations until convergence. This increased to approximately 6 hours for (3o,3o,3o) or (3n,3n,3n) clusterings. The time is consumed almost entirely in fitting in each iteration the 2398 logistic regression models for all the plants. For comparison, a single run with the multinomial model (taking approximately 8 iterations on average until convergence) takes only about 1 minute, since the multinomial model is easily fit by taking simple counts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Discussion and Future Work</head><p>Our experiments have shown that using FL-clustering we can find multiple meaningful clusterings of categorical data. The objective in our approach is explana-tory (identify underlying factors that determine the overall data patterns) rather than descriptive (provide the user with multiple views of the data).</p><p>For our purpose, it is clearly essential to use a conditional model P (X | L) of a restricted functional form, rather than an unconstrained multinomial model. Logistic regression models are a canonical choice, and can be seen as a categorical data analogue to the linear mappings between latent and observed dimensions in the factor analysis model.</p><p>A common objective in multiple clustering is that different clusterings are in some sense orthogonal or complementary. We are not yet able to say in which sense, or to what extent, FL-clustering satisfies such an objective. Empirically, a bias towards learning complementary clusterings was difficult to verify with our data, since most natural candidate segmentations based on hidden environmental variables would exhibit rather similar patterns (and not at all resemble the segmentations in Figure <ref type="figure" target="#fig_1">2</ref>). Theoretically, one can note that a multi-clustering L = l in which two clusterings are identical can not be a local maximum of the likelihood (6) (except for some degenerate, noise-free, data sets). FL-clustering, thus, is biased away from returning multiple identical clusterings. How this can be strengthened into a formal result linking likelihood gain and complementarity of different clusterings is a subject for future work.</p><p>In our experiments we have used data with a spatial structure on the data instances. Within this paper, we have used the spatial structure only for the visualization of the clustering (i.e., segmentation) results. The model can equally be used for other categorical data, and is especially suited for high-dimensional binary data (such as text document data).</p><p>On the other hand, our work was also specifically motivated by spatial data, and the relationship in this case of multiple clustering with factorial hidden Markov models <ref type="bibr" target="#b2">[3]</ref> and factorial Markov random fields <ref type="bibr" target="#b6">[7]</ref>. For spatial data one can impose a Markov random field structure on the latent variables L, i.e., the assumption of a uniform distribution for L which we used to derive <ref type="bibr" target="#b5">(6)</ref> is replaced, e.g., by the assumption that P (L = l | θ * L ) is a Gibbs distribution with fixed parameters θ * L . Learning in such a setting proceeds in the same way as described in Section 4, only that P (L = l | θ * L ) has to be added as a likelihood factor. The optimization in step ii will then usually not be possible precisely, and require an approximate solution. In this paper we did not employ a Markov random field model, since this would usually be used to enforce some smoothness and contiguity properties of the learned segments, which, for our data, seems unwarranted (considering, e.g., the rugged outline of the mountain areas).</p><p>In this paper we focused on the core of a probabilistic (multi-) clustering model, i.e., the joint distribution of latent and observable variables. In this model, the number of clusterings, and the number of clusters in each clustering is fixed. We remark, however, that either model selection techniques like BIC or MDL scoring, or a nonparametric Bayesian 'wrapper' around the core model can be used to also learn the model structure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion</head><p>We proposed a latent variable model for multiple clustering of categorical data based on a logistic regression model for the conditional distribution of the observed features. We believe that in analogy to successful techniques for dimensionality reduction, a restricted distributional form for the noisy transformation between the latent and the observed features can be instrumental for revealing relevant patterns in the latent feature space.</p><p>For clustering based on a latent variable model we have suggested a simple optimization of the conditional likelihood of the data given the latent variables, with a fixed marginal distribution for the latent variables. This leads to a learning procedure that is linear in the number of observed features, and enables us to experiment with high-dimensional biogeographical data. Our preliminary results from these experiments demonstrate the ability of the method to discover clusterings that represent meaningful explanatory factors for the data. However, further work is needed to consolidate the results returned for this data, and to investigate their potential biological meaning.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Mapping areas with mountain -valley division (a), and previous segmentation of valley areas into floristic regions [12] (b)</figDesc><graphic coords="6,309.34,128.16,161.21,104.09" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Artificial segmentations (top); "Wrong" clusterings (bottom)</figDesc><graphic coords="7,181.86,214.82,119.90,110.71" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .Fig. 4 .</head><label>34</label><figDesc>Fig. 3. Mutual information: synthetic data (a), real data (b)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. Clustering result using multinomial P (X | L)</figDesc><graphic coords="10,184.62,198.38,119.90,110.71" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">www.wsl.ch/land/products/webflora/welcome-en.ehtml</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">The data is available at http://www.wsl.ch/info/mitarbeitende/wohlgemu/lehre EN/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Categorical Data Analysis</title>
		<author>
			<persName><forename type="first">A</forename><surname>Agresti</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
			<publisher>Wiley</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Model-based multidimensional clustering of categorical data</title>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Poon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
	<note>To appear</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Factorial hidden markov models</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ghahramani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jordan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="245" to="273" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Variational inference for nonparametric multiple clustering</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Guan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Dy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Niu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ghahramani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD10 Workshop on Discovering, Summarizing and Using Multiple Clusterings</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Loglinear Models with Latent Variables</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Hagenaars</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Number 94 in Quantitative Applications in the Social Sciences</title>
				<imprint>
			<publisher>Sage Publications</publisher>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Simultaneous unsupervised learning of disparate clusterings</title>
		<author>
			<persName><forename type="first">P</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Meka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Dhillon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIAM Int. Conf. on Data Mining</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="858" to="869" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Factorial markov random fields</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zabih</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ECCV 2002, number 2352 in LNCS</title>
				<meeting>of ECCV 2002, number 2352 in LNCS</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="321" to="334" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Multiple non-redundant spectral clustering views</title>
		<author>
			<persName><forename type="first">D</forename><surname>Niu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Dy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Jordan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 27th Int. Conf. on Machine Learning (ICML-10)</title>
				<meeting>of the 27th Int. Conf. on Machine Learning (ICML-10)</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">An agglomerative method for classification of plant communities</title>
		<author>
			<persName><forename type="first">L</forename><surname>Orloci</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of Ecology</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="193" to="206" />
			<date type="published" when="1967">1967</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Bayesian cluster ensembles</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Shan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Banerjee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistical Analysis and Data Mining</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="54" to="70" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Biogeographical regionalization of switzerland based on floristic data: How many species are needed?</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wohlgemuth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biodiversity Letters</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="180" to="191" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Ein floristischer ansatz zur biogeographischen gliederung der schweiz</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wohlgemuth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Botanica Helvetica</title>
		<imprint>
			<biblScope unit="volume">106</biblScope>
			<biblScope unit="page" from="227" to="260" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Hierarchical latent class models for cluster analysis</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">L</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="697" to="723" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Discovery of latent structures: Experience with the COIL challenge 2000 data set</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Systems Science and Complexity</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="172" to="183" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
