<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Using Multi-Objective Optimization for the Selection of Ensemble Members</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Tomáš</forename><surname>Bartoň</surname></persName>
							<email>tomas.barton@fit.cvut.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Information Technology</orgName>
								<orgName type="institution">Czech Technical University in Prague</orgName>
								<address>
									<addrLine>Thakurova 9</addrLine>
									<settlement>Prague 6</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Institute of Molecular Genetics of the ASCR</orgName>
								<address>
									<addrLine>v. v. i</addrLine>
									<postCode>1083, 142 20</postCode>
									<settlement>Prague 4</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Pavel</forename><surname>Kordík</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Information Technology</orgName>
								<orgName type="institution">Czech Technical University in Prague</orgName>
								<address>
									<addrLine>Thakurova 9</addrLine>
									<settlement>Prague 6</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Using Multi-Objective Optimization for the Selection of Ensemble Members</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">A098A067A4632021D8A12414AC480249</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T16:10+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we propose a clustering process which uses a multi-objective evolution to select a set of diverse clusterings. The selected clusterings are then combined using a consensus method. This approach is compared to a clustering process where no selection is applied. We show that careful selection of input ensemble members can improve the overall quality of the final clustering. Our algorithm provides more stable clustering results and in many cases overcomes the limitations of base algorithms.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Clustering is a popular technique that can be used to reveal patterns in data or as a preprocessing step in data analysis. Clustering tends to group similar objects into groups that are called clusters and to place dissimilar objects into different clusters. Its application domains include data mining, information retrieval, bioinformatics, image processing and many others.</p><p>Many clustering algorithms have been introduced so far, an extensive overview of clustering techniques can be found in <ref type="bibr" target="#b0">[1]</ref>. Since we have the No Free Lunch Theorem <ref type="bibr" target="#b35">[36]</ref>, which states that there is no single, supreme algorithm that would optimally run on all datasets, it is complicated for a user to decide which algorithm to choose.</p><p>Ensemble methods have been successfully applied in supervised learning <ref type="bibr" target="#b24">[25]</ref> and similar attempts appeared in the unsupervised learning area <ref type="bibr" target="#b31">[32]</ref>, <ref type="bibr" target="#b13">[14]</ref>, <ref type="bibr" target="#b33">[34]</ref>. Cluster ensemble methods should eliminate the drawbacks of individual methods by combining multiple solutions into one final clustering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Cluster Ensembles</head><p>The main objective of clustering ensembles is to combine multiple clusterings into one, preferably high-quality solution.</p><p>Let X = {x 1 , x 2 , . . . , x n } be a set of n data points, where each x i ∈ X is represented by a vector of d attributes. A cluster ensemble is defined as Π = {π 1 , π 2 , . . . , π m } with m clusterings. Each base clustering π i consists of a set of clusters π i = {C i 1 ,C i 2 , . . . ,C i k i }, such that ∪ k i j=1 C i j = X, where k i is the number of clusters in a given ensemble member (it does not have to be the same for all members).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>The problem is how to obtain a final clustering π</head><formula xml:id="formula_0">* = {C * 1 ,C * 2 , . . . ,C * K },</formula><p>where K is the number of clusters in the result and π * summarizes the information from ensemble Π.</p><p>The process consists of two major steps. Firstly, we need to generate a set of clustering solutions and secondly we need ti combine the information from these solutions. A typical result of the ensemble process is a single clustering. It has been shown in supervised ensembles that the best results are achieved when using a set of predictors whose errors are dissimilar <ref type="bibr" target="#b24">[25]</ref>. Thus it is desirable to introduce diversity between ensemble members.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Ensemble Generation Strategies</head><p>Several approaches have been used to initialize clustering solutions in order to create an ensemble.</p><p>• Homogeneous ensembles: Base clusterings are created using repeated runs of a single clustering algorithm. This is quite a popular approach, repeated runs of k-means with random center initialization have been used in <ref type="bibr" target="#b13">[14]</ref>. When using k-means the number of clusters is typically fixed to √ n , where n is the size of the dataset <ref type="bibr" target="#b21">[22]</ref>.</p><p>• Varying k: Repeated runs of k-means with random initialization and k <ref type="bibr" target="#b14">[15]</ref>, a golden standard is using k in the range from 2 to √ n.</p><p>• Random subspacing An ensemble is created from base clusterings that use different initial data. This could be achieved by projecting data onto different subspaces <ref type="bibr" target="#b12">[13]</ref>, <ref type="bibr" target="#b15">[16]</ref> choosing different subsets of features <ref type="bibr" target="#b31">[32]</ref>, <ref type="bibr" target="#b36">[37]</ref>, or using data sampling techniques <ref type="bibr" target="#b9">[10]</ref>.</p><p>• Heterogeneous ensembles: Diversity of solutions is introduced by applying different algorithms on the same dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Consensus Functions</head><p>Having multiple clusterings in an ensemble, many functions have been proposed to derive a final clustering.</p><p>When only one solution is considered as the result, it is usually referred as a consensus function, unlike meta clustering where the output is a set of multiple clusterings <ref type="bibr" target="#b5">[6]</ref>.</p><p>There are several approaches as to how to represent information contained in base clusterings, some use matrices while others use graph representation.</p><p>• Pairwise similarities: A pairwise similarity matrix is created and afterwards a clustering algorithm (e.g. hierarchical agglomerative clustering) is applied to group together items that were most frequently together in the same cluster in all the base clusterings <ref type="bibr" target="#b13">[14]</ref>. the Cluster-based Similarity Partitioning Algorithm (CSPA) from Strehl and Ghosh <ref type="bibr" target="#b31">[32]</ref> uses METIS <ref type="bibr" target="#b23">[24]</ref> for partitioning a similarity matrix into k components.</p><p>• Feature-based approach: The ensemble problem is formulated as categorical data clustering. For each data point an m-dimensional vector containing labels in base clusterings is created. The goal is to find a partition π * which summarizes the information gathered from the partitions Π [28], <ref type="bibr" target="#b32">[33]</ref>, <ref type="bibr" target="#b3">[4]</ref>.</p><p>• Graph based: Many methods use graph representation for capturing relationships between base clusterings. Strehl and Ghosh <ref type="bibr" target="#b31">[32]</ref> also proposed the HyperGraph-Partitioning Algorithm (HGPA), where vertices correspond to data points and a hyperedge represents clusters. Another approach chooses CO-MUSA <ref type="bibr" target="#b26">[27]</ref> which increases the weight of the edge for each occurrence of data pairs in the same cluster. Afterwards the nodes are sorted by the attachment score, which is defined as the ratio between the sum of the node's weights and its number of incident edges. The nodes with the highest attachment score are then used as a foundation for new clusters. This approach is relatively fast to compute, however it might fail to capture complex relationships between very diverse clusterings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Multi-Objective Clustering</head><p>Multi-objective clustering usually optimizes two objective functions. Using more than a few objectives is not usual because the whole process of optimization becomes less effective.</p><p>The first multi-objective evolutionary clustering algorithm was introduced in 2004 by Handl and Knowles <ref type="bibr" target="#b17">[18]</ref> and is called VIENNA (the Voronoi Initialized Evolutionary Nearest-Neighbour Algorithm).</p><p>Subsequently, in 2007 Handl and Knowles published a Pareto-based multi-objective evolutionary clustering algorithm called MOCK <ref type="bibr" target="#b18">[19]</ref> (Multi-Objective Clustering with automatic K-determination). Each individual in MOCK is represented as a directed graph which is then translated into a clustering. The genotype is encoded as an array of integers whose length is same as the number of instances in the dataset. Each number is a pointer to another instance (an edge in the graph), since it is connected to the instance at a given index. This easily enables the application of mutation and crossover operations.</p><p>As a Multi-Objective Evolutionary Algorithm (MOEA), MOCK employs the Pareto Envelope-based Selection Algorithm version 2 (PESA-II) <ref type="bibr" target="#b6">[7]</ref>, which keeps two populations, an internal population of fixed size and a larger external population which is exploited to explore good solutions. Two complementary objectives, deviation and connectivity, are used as objectives in the evolutionary process.</p><p>A clear disadvantage of MOCK is its computation complexity, which is a typical characteristic of evolutionary algorithms. Nevertheless, the computation time spent on MOCK should result in high-quality solutions. Faceli et al. <ref type="bibr" target="#b11">[12]</ref> reported that for some high-dimensional data it is not guaranteed that the algorithm will complete, unless the control front distribution has been adjusted for the given data set.</p><p>Faceli et al. <ref type="bibr" target="#b11">[12]</ref> combined a multi-objective approach to clustering with ensemble methods and the resulting algorithm is called MOCLE (Muli-Objective Clustering Ensemble Algorithm). The objectives used in the MOCLE algorithm are the same as those used in MOCK <ref type="bibr" target="#b18">[19]</ref>: deviation and connectivity. Unlike MOCK, in this case the evolutionary algorithm used is NSGA-II <ref type="bibr" target="#b8">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Our Approach</head><p>There are many ways to produce a final consensus, nonetheless in this work we focus on the selection of highquality and diverse clusterings.</p><p>In order to optimize ensemble member selection, we apply a multi-objective optimization. The whole process is shown in Fig. <ref type="figure" target="#fig_0">1</ref>. In our previous work <ref type="bibr" target="#b2">[3]</ref> we have shown that introducing multiple objectives into the clustering process can improve the quality of clustering, specifically using the Akaike information criterion (AIC) <ref type="bibr" target="#b1">[2]</ref> (or BIC <ref type="bibr" target="#b30">[31]</ref>) with another criterion leads to better results. The second criterion is typically based on computing a ratio between cluster compactness and the sum of distances between cluster centres.</p><p>AIC is typically used in supervised learning when trying to estimate model error. Essentially it attempts to estimate the optimism of the model and then add it to the error <ref type="bibr" target="#b19">[20]</ref>:</p><formula xml:id="formula_1">f AIC = −2 • log (L n (k)) n + 2 • k n<label>(1)</label></formula><p>where L n (k) is the maximum likelihood of a model with k parameters based on a sample of size n.</p><p>The SD index was introduced in 2000 by Halkidi et al. <ref type="bibr" target="#b16">[17]</ref> and it is based on concepts of average cluster scattering and total separation of clusters which were previously used by Rezaee et al. <ref type="bibr" target="#b28">[29]</ref> for evaluation of fuzzy clusterings. The average scattering is defined as:</p><formula xml:id="formula_2">Scatt(k) = 1 k k ∑ i=1 σ (c i ) σ (X)<label>(2)</label></formula><p>where x is the norm of a vector, ci is the centroid of the i-th cluster, σ (X)</p><p>is the variance of the input dataset.</p><p>σ (X) ∈ R m with m being the number of dataset dimensions. Variance for a dimension d is defined as:</p><formula xml:id="formula_3">σ d = 1 n n ∑ i=1 x d i − xd 2 σ (X) = m ∑ i=1 (σ d ) 2</formula><p>The total separation is given by:</p><formula xml:id="formula_4">Dis(k) = D max D min k ∑ i=1 k ∑ j=1 ci − c j −1<label>(3)</label></formula><p>where D max is the maximum distance and D min is the minimum distance between cluster centers ( ci ) and k is the number of clusters.</p><formula xml:id="formula_5">D max = max i, j∈{1,...,k} i = j ( ci − c j )<label>(4)</label></formula><formula xml:id="formula_6">D min = min i, j∈{1,...,k} i = j ( ci − c j )<label>(5)</label></formula><p>Then we can define the SD validity index as follows:</p><formula xml:id="formula_7">f SD = α • Scat(k) + Dis(k)<label>(6)</label></formula><p>where should be a weighting factor equal to Dis(c max ) with c max being the maximum number of clusters <ref type="bibr" target="#b16">[17]</ref>. This makes perfect sense for fuzzy clustering (as was proposed in <ref type="bibr" target="#b28">[29]</ref>), however it is rather unclear how to compute c max in the case of crisp clustering, when c max k without running another clustering with c max as the requested number of clusters. Nonetheless, <ref type="bibr" target="#b16">[17]</ref> mentions that "SD proposes an optimal number of clusters almost irrespective of c max , the maximum number of clusters", thus we consider the special case where c max = k:</p><formula xml:id="formula_8">f SD = Dis(k) • Scat(k) + Dis(k) (7) = Dis(k) • (Scat(k) + 1)<label>(8)</label></formula><p>The idea of minimizing inner cluster distances and maximizing distances between cluster centres it not really new. In fact most of the clustering objective functions use this idea. Instead of the SD index we could have used C-index <ref type="bibr" target="#b20">[21]</ref>, Calinski-Harabasz Index <ref type="bibr" target="#b4">[5]</ref>, Davies-Bouldin <ref type="bibr" target="#b7">[8]</ref> or Dunn's index <ref type="bibr" target="#b10">[11]</ref>, just to name a few. Each of these indexes might perform better on some dataset, however it is out of scope of this paper to compare which combination would be better. We considered</p><p>As an evolutionary algorithm we use NSGA-II <ref type="bibr" target="#b8">[9]</ref> while using only the mutation operator and a relatively small population (see Table <ref type="table" target="#tab_1">2</ref>). Each individual in the population contains a configuration of a clustering algorithm. For this experiment we used only the basic k-means <ref type="bibr" target="#b25">[26]</ref> algorithm with random initialization. The number of clusters k is the only parameter that we change during the evolutionary process. For k we used a constraint in order to keep the number of clusters within reasonable boundaries. For all test datasets we used the interval 2, √ n . In order to evaluate the effect of Pareto optimal solutions selection for the ensemble bagging process, we compare two strategies for generating base clusterings. The first one generates 10 k-means clusterings with random initialization with k within the interval 2, √ n . The second method uses multi-objective evolution of clusterings where each individual represents a k-means clustering.</p><p>As a consensus function we use either a graph based COMUSA algorithm or a hierarchical agglomerative clustering of the co-association matrix. Both approaches have their drawbacks, the COMUSA algorithm is locally consistent, however fails to integrate the information contained in different base clusterings, thus final clustering does not overcome limitations of base clustering algorithms. The latter could be very slow on larger datasets (complexity O(n 3 )) and might produce noisy clustering in case of very diverse base clusterings. To evaluate clustering quality we use NMI 1 (Normalized Mutual Information) <ref type="bibr" target="#b31">[32]</ref>. NMI has proved to be a better criterion than the Adjusted Rand Index (ARI) for the evaluation of datasets with apriori known labels. However it does not capture noise in clusters (which does not occur in the case of traditional k-means), which might be introduced by some consensus methods. Another limitation of these evaluation metrics is apparent from Figure <ref type="figure" target="#fig_5">6</ref>, where clustering with obvious 50% assignment error gets lowest possible score. Both indexes do not penalize correctly higher number of clusters, thus algorithms producing many small clusters will be preferred by these objectives.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>Most of the datasets used in these experiments come from the Fundamental Clustering Problems Suite <ref type="bibr" target="#b34">[35]</ref>. We have intentionally chosen low dimensional data in order to be able to visually evaluate the results. An overview of the datasets used can be found in Table <ref type="table" target="#tab_0">1</ref>.</p><p>As objectives during the multi-objective evolution we used AIC and SD index in all cases. The advantage is combination of two measures that are based on very different grounds. We tried using Deviation and Connectivity, as it was proposed by Handl <ref type="bibr" target="#b18">[19]</ref> (and later used by <ref type="bibr" target="#b11">[12]</ref>), but for all tested datasets, our approach was better or at least comparable with these objectives.</p><p>In the case of the COMUSA approach, the multiobjective selection either improved the final NMI score or it was comparable with the random selection. There were only a few exceptions, especially in case of compact datasets like the 9 diamond dataset which contains a specific regular pattern that is unlikely to appear in any realworld dataset (see Table <ref type="table" target="#tab_3">3</ref>). The number of clusters produces by COMUSA approach was usually higher than the correct number of clusters and clustering contains many nonsense clusters. Despite these facts the NMI score does not penalize these properties appropriately.</p><p>Agglomerative clustering of the co-association matrix despite its high computational requirement provide pretty good results. Nonetheless in the case of the chainlink dataset there is over 50% improvement in NMI. It is important to note, that for HAC-RAND and HAC-MO we did not provide the information about the correct number of clusters. Still these methods manages to estimate number of clusters or at least provide result that is very close to the number of supervised classes.  <ref type="table">4</ref>: NMI sqrt score from the clustering of various artificial datasets. As a consensus method, hierarchical agglomerative clustering (with complete linkage) of coassociation matrix is used. In the case of HAC-RAND we run 10 independent k-means clusterings with a random number of clusters (between 2 and √ n), then form a co-association matrix and finally run agglomerative clustering of the matrix. HAC-RAND works in very similar manner, but instead of the first step a multi-objective evolution of k-means is performed. 10 dominating solutions are selected and the rest of the algorithm is the same. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>During our experiments we have shown that careful selection of clusterings for the ensemble process can significantly improve overall clustering quality for non-trivial datasets (measured by NMI).</p><p>It is interesting that non-spherical clusters could be discovered by consensus function when agglomerative hierarchical clustering is used (compare Fig. <ref type="figure" target="#fig_2">2 and 3</ref>).</p><p>Using a multi-objective optimization for clustering se-  lection improved the overall quality of clustering results, however ensemble methods might produce noisy clustering with a higher evaluation score. Noisy clusterings are hard to measure with current evaluation metrics, therefore it might be beneficial to include an unsupervised score in the results. In further research we would like to examine the process of selecting optimal objectives for each dataset.  </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Cluster Ensemble Process: Firstly using the same input dataset we generate multiple clusterings, then select m best clusterings for the ensemble and combine them into a single solution.</figDesc><graphic coords="3,56.76,68.46,218.33,283.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Typical clustering longsquare dataset using k-means (k = 6). K-Means algorithm fails to reveal nonspherical clusters (Ajusted Rand Index = 0.94, NMI = 0.85).</figDesc><graphic coords="5,56.76,414.80,218.32,157.73" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Consensus of 10 independent k-means runs on the longsquare dataset using k = 6. The co-association matrix obtained was clustered using hierarchical clustering (average linkage). The resulting clustering is qualitatively better than a single k-means run, however it contains noise in all clusters (Ajusted Rand Index = 0.99, NMI = 0.95).</figDesc><graphic coords="5,320.19,344.55,218.33,155.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Consensus of 10 independent k-means runs on the long1 dataset using k = 6. On many datasets setting the correct k does not improve the quality of the resulting clustering. The co-association matrix obtained was clustered using hierarchical clustering (average linkage).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Consensus of 10 independent k-means runs on the triangle1 dataset using k = 6. In this case the consensus produces a nonsense cluster that mixes items from 3 other clusters together.</figDesc><graphic coords="6,56.76,68.46,218.32,141.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Single k-means runs on the long1 dataset using k = 2. K-Means algorithm fails to reveal non-spherical clusters. Both supervised indexes Ajusted Rand Index and NMI assigns this clustering score 0.0, even though there is 50% assignment error.</figDesc><graphic coords="6,56.76,302.47,218.32,124.68" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Datasets used for experiments.</figDesc><table><row><cell>Dataset</cell><cell>d</cell><cell cols="3">n classes source</cell></row><row><cell cols="2">aggregation 2</cell><cell>788</cell><cell>7</cell><cell>[15]</cell></row><row><cell>atom</cell><cell>3</cell><cell>800</cell><cell>2</cell><cell>[35]</cell></row><row><cell>chainlink</cell><cell cols="2">3 1000</cell><cell>2</cell><cell>[35]</cell></row><row><cell>complex8</cell><cell cols="2">2 2551</cell><cell>8</cell><cell>[30]</cell></row><row><cell>complex9</cell><cell cols="2">2 3031</cell><cell>9</cell><cell>[30]</cell></row><row><cell>diamond9</cell><cell cols="2">2 3000</cell><cell>9</cell><cell>[30]</cell></row><row><cell>jain</cell><cell>2</cell><cell>373</cell><cell>2</cell><cell>[23]</cell></row><row><cell>long1</cell><cell cols="2">2 1000</cell><cell>2</cell><cell>[18]</cell></row><row><cell>longsquare</cell><cell>2</cell><cell>900</cell><cell>6</cell><cell>[18]</cell></row><row><cell>lsun</cell><cell>2</cell><cell>400</cell><cell>3</cell><cell>[35]</cell></row><row><cell>target</cell><cell>2</cell><cell>770</cell><cell>6</cell><cell>[35]</cell></row><row><cell>tetra</cell><cell>3</cell><cell>400</cell><cell>4</cell><cell>[35]</cell></row><row><cell>triangle1</cell><cell cols="2">2 1000</cell><cell>4</cell><cell>[18]</cell></row><row><cell>Parameter</cell><cell></cell><cell cols="2">Setting</cell><cell></cell></row><row><cell cols="2">Number of generations</cell><cell>5</cell><cell></cell><cell></cell></row><row><cell>Population size</cell><cell></cell><cell>10</cell><cell></cell><cell></cell></row><row><cell cols="2">Mutation probability</cell><cell>30%</cell><cell></cell><cell></cell></row><row><cell>Mutation</cell><cell></cell><cell cols="3">Polynomial mutation</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Parameter settings for NSGA-II evolution</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>1</head><label></label><figDesc>There are several versions of this evaluation method, sometimes it is referred as NMI sqrt</figDesc><table><row><cell>Dataset</cell><cell cols="3">C-RAND C-MO k-means</cell></row><row><cell>aggregation</cell><cell>0.75</cell><cell>0.84</cell><cell>0.84</cell></row><row><cell>atom</cell><cell>0.39</cell><cell>0.61</cell><cell>0.28</cell></row><row><cell>chainlink</cell><cell>0.07</cell><cell>0.50</cell><cell>0.07</cell></row><row><cell>complex8</cell><cell>0.59</cell><cell>0.64</cell><cell>0.59</cell></row><row><cell>complex9</cell><cell>0.64</cell><cell>0.66</cell><cell>0.65</cell></row><row><cell>diamond9</cell><cell>1.00</cell><cell>0.87</cell><cell>0.97</cell></row><row><cell>jain</cell><cell>0.36</cell><cell>0.51</cell><cell>0.37</cell></row><row><cell>long1</cell><cell>0.00</cell><cell>0.49</cell><cell>0.03</cell></row><row><cell>longsquare</cell><cell>0.83</cell><cell>0.86</cell><cell>0.80</cell></row><row><cell>lsun</cell><cell>0.54</cell><cell>0.69</cell><cell>0.54</cell></row><row><cell>target</cell><cell>0.67</cell><cell>0.57</cell><cell>0.69</cell></row><row><cell>tetra</cell><cell>1.00</cell><cell>0.86</cell><cell>0.99</cell></row><row><cell>triangle1</cell><cell>0.95</cell><cell>0.78</cell><cell>0.94</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 :</head><label>3</label><figDesc>NMI sqrt score from various artificial dataset clusterings. C-RAND uses bagging of 10 k-means runs with a fixed value k that corresponds to the number of classes in the given dataset. The same approach is used in the case of k-means, but without bagging. C-MO is a NSGA-II based algorithm with criteria AIC and SD index; the number of clusters is randomized during evolution within the interval 2 to √ n. Both C-RAND and C-MO use the COMUSA algorithm to form final (consensus) partitioning. The NMI values are averages from 30 independent runs.</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Using Multi-Objective Optimization for the Selection of Ensemble Members</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgement</head><p>The research reported in this paper was supported by the Czech Science Foundation (GA ČR) grant 13-17187S.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Data clustering: algorithms and applications</title>
		<editor>Aggarwal, C. C., Reddy, C. K.</editor>
		<imprint>
			<date type="published" when="2014">2014</date>
			<publisher>CRC Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A new look at the statistical model identification</title>
		<author>
			<persName><forename type="first">H</forename><surname>Akaike</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Automatic Control</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="716" to="723" />
			<date type="published" when="1974-12">December 1974</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Evaluation of relative indexes for multi-objective clustering</title>
		<author>
			<persName><forename type="first">T</forename><surname>Bartoň</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kordík</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Hybrid Artificial Intelligent Systems</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">E</forename><surname>Onieva</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Santos</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Osaba</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Quintián</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Corchado</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="volume">9121</biblScope>
			<biblScope unit="page" from="465" to="476" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Combining multiple clustering systems</title>
		<author>
			<persName><forename type="first">C</forename><surname>Boulis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ostendorf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PKDD</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J.-F</forename><surname>Boulicaut</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Esposito</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Giannotti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Pedreschi</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="volume">3202</biblScope>
			<biblScope unit="page" from="63" to="74" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A dendrite method for cluster analysis</title>
		<author>
			<persName><forename type="first">T</forename><surname>Caliński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Harabasz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications in Statistics-theory and Methods</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="27" />
			<date type="published" when="1974">1974</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Meta clustering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Caruana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Elhawary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Smith</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Sixth International Conference on Data Mining</title>
				<meeting>the Sixth International Conference on Data Mining<address><addrLine>Washington, DC, USA</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="107" to="118" />
		</imprint>
	</monogr>
	<note>ICDM&apos;06</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">PESA-II: region-based selection in evolutionary multiobjective optimization</title>
		<author>
			<persName><forename type="first">D</forename><surname>Corne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Jerram</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Knowles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Oates</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Genetic and Evolutionary Computation Conference</title>
				<meeting>the Genetic and Evolutionary Computation Conference</meeting>
		<imprint>
			<publisher>GECCO</publisher>
			<date type="published" when="2001">2001. 2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A cluster separation measure</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Davies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Bouldin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Pattern Anal. Mach. Intell</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="224" to="227" />
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A fast and elitist multiobjective genetic algorithm: NSGA-II</title>
		<author>
			<persName><forename type="first">K</forename><surname>Deb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pratap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Meyarivan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Evolutionary Computation</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="182" to="197" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Bagging to improve the accuracy of a clustering procedure</title>
		<author>
			<persName><forename type="first">S</forename><surname>Dudoit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fridlyand</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="1090" to="1099" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Well-separated clusters and optimal fuzzy partitions</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Dunn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Cybernetics</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="95" to="104" />
			<date type="published" when="1974">1974</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Multi-objective clustering ensemble for gene expression data analysis</title>
		<author>
			<persName><forename type="first">K</forename><surname>Faceli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C P</forename><surname>De Souto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S A</forename><surname>De Araujo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C P L F</forename><surname>De Carvalho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">72</biblScope>
			<biblScope unit="issue">13-15</biblScope>
			<biblScope unit="page" from="2763" to="2774" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Random projection for high dimensional data clustering: a cluster ensemble approach</title>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">Z</forename><surname>Fern</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">E</forename><surname>Brodley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICML</title>
				<editor>
			<persName><forename type="first">T</forename><surname>Fawcett</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Mishra</surname></persName>
		</editor>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="186" to="193" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Combining multiple clusterings using evidence accumulation</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L N</forename><surname>Fred</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Jain</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Pattern Anal. Mach. Intell</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="835" to="850" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Clustering aggregation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gionis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mannila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tsaparas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">TKDD</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Projective clustering ensembles</title>
		<author>
			<persName><forename type="first">F</forename><surname>Gullo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Domeniconi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tagarelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Min. Knowl. Discov</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="452" to="511" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Quality scheme assessment in the clustering process</title>
		<author>
			<persName><forename type="first">M</forename><surname>Halkidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vazirgiannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Batistakis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PKDD</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Zighed</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Komorowski</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zytkow</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2000">2000</date>
			<biblScope unit="volume">1910</biblScope>
			<biblScope unit="page" from="265" to="276" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Evolutionary multiobjective clustering</title>
		<author>
			<persName><forename type="first">J</forename><surname>Handl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Knowles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Parallel Problem Solving from Nature -PPSN VIII</title>
				<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="1081" to="1091" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">An evolutionary approach to multiobjective clustering</title>
		<author>
			<persName><forename type="first">J</forename><surname>Handl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Knowles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Evolutionary Computation</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="56" to="76" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
	<note>IEEE Transactions on</note>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">The elements of statistical learning</title>
		<author>
			<persName><forename type="first">T</forename><surname>Hastie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tibshirani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Corporation</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>Springer</publisher>
			<pubPlace>Dordrecht</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A general statistical framework for assessing categorical clustering in free recall</title>
		<author>
			<persName><forename type="first">L</forename><surname>Hubert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Levin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychological Bulletin</title>
		<imprint>
			<biblScope unit="volume">83</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">1072</biblScope>
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Improved link-based cluster ensembles</title>
		<author>
			<persName><forename type="first">N</forename><surname>Iam-On</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Boongoen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IJCNN (2012), IEEE</title>
				<imprint>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Data clustering: a user&apos;s dilemma</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">H C</forename><surname>Law</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>S. K. Pal, S. Bandyopadhyay, and S. Biswas</editor>
		<imprint>
			<biblScope unit="volume">3776</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2005">2005</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A fast and high quality multilevel scheme for partitioning irregular graphs</title>
		<author>
			<persName><forename type="first">G</forename><surname>Karypis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kumar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIAM J. Sci. Comput</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="359" to="392" />
			<date type="published" when="1998-12">December 1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Combining classifiers</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kittler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hatef</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">P W</forename><surname>Duin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Sixth International Conference on Pattern Recognition</title>
				<meeting>the Sixth International Conference on Pattern Recognition<address><addrLine>Silver Spring, MD</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society Press</publisher>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="897" to="901" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Some methods for classification and analysis of multivariate observations</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Macqueen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Fifth Berkeley Symposium on Mathematical Statistics and Probability</title>
				<editor>
			<persName><forename type="first">L</forename><forename type="middle">M L</forename><surname>Cam</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Neyman</surname></persName>
		</editor>
		<meeting>of the Fifth Berkeley Symposium on Mathematical Statistics and Probability</meeting>
		<imprint>
			<publisher>University of California Press</publisher>
			<date type="published" when="1967">1967</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="281" to="297" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Obtaining better quality final clustering by merging a collection of clusterings</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mimaroglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Erdil</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">20</biblScope>
			<biblScope unit="page" from="2645" to="2646" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Consensus clusterings</title>
		<author>
			<persName><forename type="first">N</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caruana</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Computer Society</title>
				<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="607" to="612" />
		</imprint>
	</monogr>
	<note>ICDM</note>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">A new cluster validity index for the fuzzy c-mean</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Rezaee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lelieveldt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reiber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">3-4</biblScope>
			<biblScope unit="page" from="237" to="246" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Determining the number of clusters/segments in hierarchical clustering/segmentation algorithms</title>
		<author>
			<persName><forename type="first">S</forename><surname>Salvador</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Chan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICTAI</title>
				<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="576" to="584" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Estimating the dimension of a model</title>
		<author>
			<persName><forename type="first">G</forename><surname>Schwarz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annals of Statistics</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="461" to="464" />
			<date type="published" when="1978">1978</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Cluster ensembles -a knowledge reuse framework for combining multiple partitions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Strehl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ghosh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal on Machine Learning Research (JMLR)</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="583" to="617" />
			<date type="published" when="2002-12">December 2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Clustering ensembles: models of consensus and weak partitions</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Topchy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">F</forename><surname>Punch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Pattern Anal. Mach. Intell</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1866" to="1881" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Ensemble clustering with voting active clusters</title>
		<author>
			<persName><forename type="first">K</forename><surname>Tumer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Agogino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">14</biblScope>
			<biblScope unit="page" from="1947" to="1953" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<title level="m" type="main">ESOM-Maps: tools for clustering, visualization, and classification with Emergent SOM</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ultsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Mörchen</surname></persName>
		</author>
		<idno>No. 46</idno>
		<imprint>
			<date type="published" when="2005">2005</date>
			<pubPlace>Germany</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Dept. of Mathematics and Computer Science, University of Marburg</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">No free lunch theorem for optimization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Wolpert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">G</forename><surname>Macready</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Evolutionary Computation</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Graph-based consensus clustering for class discovery from gene expression data</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H. -S</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H. -Q</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="issue">21</biblScope>
			<biblScope unit="page" from="2888" to="2896" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
