<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A study of SOM clustering software implementations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>CCS Concepts</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>A. B. Adeyemo Computer Science Department University of Ibadan Nigeria</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>7</fpage>
      <lpage>9</lpage>
      <abstract>
        <p>Clustering algorithms generally suffer from some well-known problems for which the Self Organizing Maps (SOM) algorithms are adept at handling. While there are many variants of the SOM algorithm, software programmes that implement the SOM algorithms have tended to give varying results even when tested on the same data sets. This can have serious implications when the goal of the clustering is novelty detection. In this study a comparison of the performance of some SOM clustering software was carried out and results presented.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Comparative Analysis</kwd>
        <kwd>Clustering</kwd>
        <kwd>Self Organizing Maps</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        In the clustering process data is grouped in such a way that the
intra-cluster similarity is maximized while the inter-cluster
similarity is minimized. Data can be described by either
categorical or numeric features. Due to the differences in the
characteristics of these two kinds of data, attempts to develop
criteria functions for mixed data have not been very successful
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. There are two widely used clustering methods: the
hierarchical and the nonhierarchical (partitional) methods. The
hierarchical clustering process can be categorized as divisive
when a large data set is divided into several small groups and,
agglomerative when a small data set are put together to create
a larger cluster. Self-Organizing Maps (SOM) are competitive
networks that provide a "topological" mapping from the input
space to the clusters [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The SOM was inspired by the way in
which various human sensory impressions are neurologically
mapped into the brain such that spatial or other relations
among stimuli correspond to spatial relations among the
neurons.
      </p>
      <p>
        In a SOM, the neurons (clusters) are organized into a grid
which is usually two-dimensional, but sometimes
onedimensional (or (rarely) three or more-dimensions. The reason
for using one- and two dimensional grids is that space
structures of higher dimensionality cause problems with data
display and cannot be displayed on the monitor. The SOM
working algorithm is a variant of multidimensional vectors
clustering of which the Kmeans clustering algorithm is an
example of this type of algorithm [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        The SOM neural network uses a competitive learning algorithm
and is a method for unsupervised learning, based on a grid of
artificial neurons whose weights are adapted to match input
vectors in a training set. The SOM algorithm is fed with
feature vectors, which can be of any dimension. The algorithm
for the training of the SOM [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is explained easily in terms of
a set of artificial neurons, each having its own physical
location on the output map, which take part in a
winner-takeall process where a node with its weight vector closest to the
vector of inputs is declared the winner and its weights are
adjusted making them closer to the input vector.
In each training step, one sample vector „x‟ from the input data
set is chosen randomly and a similarity measure is calculated
between it and all the weight vectors of the map. The
BestMatching Unit (BMU), denoted as „c‟, is the unit whose
weight vector has the greatest similarity with the input sample
„x‟ (figure 1). The similarity is usually defined by means of a
distance measure, usually the Euclidian distance. The BMU is
defined mathematically as the processing element for which
the expression:
. …..…………….….. 1
where d is the distance measure.
      </p>
      <p>Each node has a set of neighbors. When a node wins a
competition, the neighbor‟s weights are also changed but not
as much as that of the winning node. The further the neighbor
is from the winner, the smaller its weight change. The SOM
update rule for the weight vector of the unit i is given
mathematically as:
………………2
where
t represents the sample index for each presentation of a sample „x‟
hc(x),i represents the neighborhood function around the winner unit
„c‟, with neighborhood radius r(t).</p>
      <p>
        The neighborhood function is like a smoothing kernel that is
time-variable. It is a decreasing function of the distance
between the the ith and cth reference vectors on the map grid.
The neighborhood function is usually expressed as the
Gaussian function which can be expressed mathematically as:
…………………3
where
ά(t) represents the learning rate factor and takes values 0&lt; ά(t)&lt;1
σ(t) represents the width of the neighborhood function which
decreases monotomically with the regression steps.
A simpler definition of the neighbourhood function given by
Kohonen [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is:
hc(x),I=σ(t)…………………………………………………….4
If ║ri – rc║ is smaller than a given radius around node „c‟ and
the radius is also a monotomically decreasing function of the
regression steps, but otherwise hc(x),I = 0. σ(t) is a
diminishing function of time. At the beginning of the learning
procedure it is fairly large, but it is made to gradually shrink
during learning. Towards the end of learning a single winning
processing element is trained. A linear diminishing function of
time is usually used. The learning process consisting of winner
selection by Equation (1) and adaptation of the synaptic
weights by Equation (2). This process is repeated for each
input vector, usually for a large number of cycles with
different inputs producing different winners. The network
therefore associates output nodes with groups or patterns in the
input data set. The SOM algorithm is very simple and allows
for many subtle adaptations.
      </p>
      <p>
        There are some visual displays that are used to "determine" where
the natural cluster boundaries are in the SOM. Some of the visual
tools that can be used are Histograms [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], Component Plane
displays [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], U-matrix, P-matrix and U* matrix displays [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ],
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], [12, [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. An important concept in interpreting these displays
is the interaction of the two properties of the SOM. These are the
neighborhood relationship, and the density mapping. Neighboring
neurons in the SOM cannot be too far away from each other (in
order to maintain their similarity) but the SOM also wants to place
more neurons in areas of high input density (for example, logical
clusters). Because of this, there will be neurons that will be placed
in areas between natural clusters which are typically low input
density areas (so that the map can "stretch" between clusters).
The standard SOM algorithm uses numeric type variables and the
Euclidean distance function. The arithmetic operations used
during the learning phase for the update of the feature vectors
cannot be used with categorical values. The SOM was not directly
designed to work with categorical variables due to the limitation
of learning laws. The method usually adopted is to translate
categories to numeric numbers during data pre-processing before
training using the transformed data using standard SOM algorithm
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The Kohonen SOM clustering algorithm has also been used
for classification purposes with remarkable results. There is a
fundamental difference between the clustering process and the
classification process. Clustering is an unsupervised process while
classification is supervised. Usually data clustering is used as a
pre-processor for classification purposes [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        A rich variety of versions of the basic SOM algorithm have
been proposed. Some of the variants aim at improving the
preservation of topology by using more flexible map structures
instead of the fixed grid. Some of these methods however
cannot be used for visualization as easily as the regular grid.
Some variants aim at reducing the computational complexity
of the SOM [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Experiments using different distance
measures, map topologies, training parameters such as the
learning rate and neighbourhood function can be carried out.
Using identical settings, training of a SOM map over different
iterations can lead to different mappings, because of the
random initialisation. Yet it has been shown that the
conclusions drawn from the map remain remarkably
consistent, which makes it a very useful tool in many different
circumstances [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Some of the desirable features that good
SOM clustering software should have include:
      </p>
      <p>Being able to set the neighborhood kernel function and
to set the start value for the neighborhood function
(learning radius): The neighborhood function
determines how strongly the processing elements are
connected to each other. Neighborhoods of different
sizes in different neuron configurations (e.g.
rectangular and hexagonal lattices) can be used. The
simplest neighborhood function is the bubble
(winnertakes-all): it is constant (or 1) over the whole
neighborhood of the winner unit and zero elsewhere.
Usually the neighbourhood function is expressed as a
Gaussian function and as expected using the
winnertakes-all function retrieves less clusters than the
Gaussian function.</p>
      <p>Being able to set the activation function and weight
initialization methods: Before the training, initial values
are given to the prototype vectors of the SOM. The
SOM is very robust with respect to the initialization
process, however, when properly accomplished it
allows the algorithm to converge faster to a good</p>
      <p>solution. Initialization procedures that have been used
are: Random initialization, where the weight vectors
are initialized with small random values; Sample
initialization, where the weight vectors are initialized
with random samples drawn from the input data set;
Linear initialization, where the weight vectors are
initialized in an orderly fashion along the linear
subspace spanned by the two principal eigenvectors of
the input data set.</p>
      <p>Being able to set the choice of cooling strategy during
training: for example linear or exponential.</p>
      <p>Being able to set the distance measure to be used, for
example, Euclidean, Manhattan and Maximum value: It
is noted that the distance measure between data points
is an important component of a clustering algorithm. If
the components of the data instance vectors are all in
the same physical units then it is possible to use the
simple Euclidean distance metric to successfully group
similar data elements. The Euclidean distance in a two
or three-dimensional space measures is the actual
geometric distance between objects in the space.
However, it has been observed that even the Euclidean
distance can sometimes be misleading, because of the
way the mathematical formula used to combine the
distances between the single components of the data
feature vectors into a unique distance measure that can
be used for clustering purposes is computed. Different
formulas lead to different clustering‟s. Therefore,
domain knowledge must be used to guide the
formulation of a suitable distance measure for each
particular application.</p>
      <p>
        Being able to set the scaling technique to be used: for
example z-transform, (0,1) transform, (1,-1) transform
or none, depending on the clustering goal and data set.
Being able to set the starting and stopping learning rate:
The learning rate is a decreasing function of time
between [
        <xref ref-type="bibr" rid="ref1">0,1</xref>
        ]. The learning rate can be expressed as a
linear function and as a function inversely proportional
to time. Using the inverse function ensures that all
input samples have approximately equal influence on
the training result. Some learning rate functions that
have been implemented are the linear, inverse-of-time,
and as a power ser.
      </p>
      <p>
        Being able to set the training algorithm to be used: for
example batch, on-line, hybrid etc. The batch algorithm
has been shown to be faster [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] than the normal
sequential algorithm (and the results are just as good or
even better).
      </p>
      <p>Good data visualization options: for example
histograms, hinton charts, weight charts (maps),
UMatrix, P-Matrix etc. Good result analysis and
presentation functions: computation of vital statistics
for evaluating the quality of the clustering for example,
mean, standard deviation (or variance), correlation
coefficient, t-test etc.</p>
      <p>This work presents a comparative study of the performance
some SOM clustering software when tested on the same data
set. Results were presented and reasons for the observed
variations presented. The study also presents the desirable
features that standard SOM software should have.</p>
    </sec>
    <sec id="sec-2">
      <title>2. MATERIALS AND METHODS</title>
      <p>Agro metrological data for FRIN headquarters, Ibadan,
Nigeria was used. The data set had 254 records and the
attributes in the data set were: Year (numeric), Month (text),
Total Rainfall in millimeters (numeric), Minimum
Temperature in Celsius (numeric), Maximum Temperature in
Celsius (numeric), Relative Humidity and Fire Danger Index
(numeric). The SOM software used were: NNClust, Pittnet
Neural Network Educational Software and RapidMiner Studio.
The NNClust software was programmed to use only the
Gaussian neighbourhood function and the Euclidean distance
measure. The user can input the learning rate and starting
neighbourhood size. The software automatically normalizes
the input data between -1 and 1 and has features for generating
data/result statistics and data visualization such as weight
maps and radar charts. The Pittnet software also uses the
Gaussian neighbourhood function and Euclidean distance
metrics. The user also defines the starting learning rate and it
also automatically normalizes the data between 0 and 1. It is a
DOS based program that saves its result in a text file and has
no data analysis or data visualization ability. RapidMiner
studio (Community Edition) has facilities for selecting
parameters for defining the learning rate, neighbourhood
radius and can choose either to normalize the data or not. It
also has an array of tools for statistical data analysis and data
visualization.</p>
      <p>Using the three software‟s clusters was generated. The
arithmetic mean of each cluster group was also computed. The
arithmetic mean is a measure of central tendency which
describes the central location of data and is usually used with
other statistical measures such as the standard deviation
because it can be affected by extreme values in the data set and
therefore be biased. The standard deviation describes the
spread of the data and is a popular measure of dispersion. It
measures the average distance between a single observation
and its mean.</p>
    </sec>
    <sec id="sec-3">
      <title>3. RESULTS AND DISCUSSION</title>
      <p>The meteorological data was clustered using NNClust SOM
clustering software with a starting learning rate of 0.9 and was
trained over 100 epochs. The software accepts only numeric
values. Non numeric values are treated as missing values
which are replaced by the column mean. The software was set
to identify a maximum of ten clusters, however only eight
clusters were generated. The software uses the number of
clusters specified to create the SOM grid. The mean and
standard deviation of the eight clusters were computed.
Increasing the training cycle did not improve the results. Table
1 presents the summary of the eight clusters, while figure 2
presents the chart of the cluster means.
The meteorological data was trained using the Pittnet software
with a starting learning rate of 0.9 and was set to train for 100
epochs, although the software stops training as soon as the
maximum number of clusters have been generated. The
software requires the user to specify the number of clusters
expected apriori. This number is used in conjunction with the
number of input signals (attributes) to determine the SOM grid
size. Expected number of clusters was set to ten. The software
identified only four clusters. The mean and standard deviation
of the clusters were computed. Table 2 presents the summary
of the clusters, while figure 3 presents the chart of the cluster
means.</p>
      <p>TheRapidMiner Studio software was used to cluster the
meteorological data set using a starting learning rate of 0.9 and
was trained over a 100 epochs. The expected number of
clusters was set at ten and the software generated ten clusters.
Table 3 presents the summary of the cluster means with their
standard deviations while figure 4 presents a chart of their
cluster means.
3.1</p>
    </sec>
    <sec id="sec-4">
      <title>Discussion of Results</title>
      <p>The quality of the clusters identified in the data by the three
software‟s can be inferred from a comparison of the mean and
standard deviation of the clusters. If the value of the standard
deviation is low, then the clustered records are within the same
range. However if the value is high this suggests the presence
of outliers in the clustered data records. For example table 4
presents the clustered records for cluster 2 (table1) for the
NNClust software which is representative of the trend
observed in the clusters identified by the software. Interpreting
the cluster is indecisive when the values in the Total Rainfall
fields are considered. The field has a mean of 142.05 and a
standard deviation of 136.011711.</p>
      <sec id="sec-4-1">
        <title>TotalRainfall</title>
      </sec>
      <sec id="sec-4-2">
        <title>MaxTemp</title>
      </sec>
      <sec id="sec-4-3">
        <title>MinTemp RH FireDangerInd ex</title>
        <p>Similarly considering the clusters identified by the Pittnet
software in table 2 the same trend is observed. Table 5
presents the records for cluster 4 (table 2) for the Pittnet
software cluster results. It can be observed that the cluster is
consists of data records which have the same value for the
FireDangerIndex attribute. However, considering the Total
Rainfall field which has a mean value of 39.74444 and a
standard deviation of 43.34732. The high standard deviation
value implies that there are outlier data values in the clustered
records.</p>
        <p>
          The clusters identified by the RapidMiner software presented
in table 3 were easier to interpret. They followed the expected
rainfall pattern which is known for the region where the data
was collected [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. Cluster 2 (table 3) contained records with
only a high FireDangerIndex of 4 as presented in table 6, while
cluster 5 (table 3) contains records with the highest recorded
Rainfall level in the data set. The other clusters also contained
data records which can be categorized by the Rainfall level
pattern of the region.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. ACKNOWLEDGMENTS</title>
      <p>Some of the problems found in the literature about clustering
algorithms are: Most clustering techniques are based on
distance calculations which are very sensitive to ranges of
variables, therefore the values have to be normalized.
Normalization however is a subjective function, and these
transformations cannot be carried out without creating biases;
The presence of outliers in data sets create problems in data
clustering based on distance calculations when they have not
been identified and removed from the data set; Handling
categorical variables (non-numeric data, non-numeric
variables, categorical data, nominal data, or nominal variables)
are a problem for most clustering algorithms, and even when
data encoding methods are used they can introduce extra
biases due to the number of values which the encoding
introduces in the categorical variables; The selection of
variables also has a large influence on clustering results, while
the assigning of different weights for variables and categorical
values can be used, when many variables and categorical
values are involved, it can affect the clustering quality;
Capturing patterns (or behaviors) hidden inside time-varying
variables and modeling them is another problem and most
clustering techniques do not possess this predictive modeling
capability; Most clustering techniques were developed for
laboratory generated simple data sets consisting of a few to
several numerical variables; hence they can‟t be used for large
data analyses that consist of many categorical complex data.
Most common implementation of data clustering algorithms
suffer from these problems, however, SOM‟s are very robust
and are adept at handling these problems but this depends also
on the goal of the algorithm‟s implementation (programming).
Applications programmed for demonstration purposes cannot
be used for large scale projects and some implementations are
not flexible and do not give users much options. However if
the various implementations of the conventional SOM
algorithm (which are usually focused on the goals of the
programmer) provides enough options to the user, it is still a
very robust algorithm that can be used for both numerical,
categorical and mixed data sets. Further work in this study is
focused on the development of an open flexible SOM
clustering tool with adequate features that can be used for
research purposes.
Portuguese conference on progress in Artificial Intelligence , pp
304 - 313, (Sringer-Verlag Berlin, Heidelberg ©2005)
Cluster 1</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]. Chang
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Ding</surname>
          </string-name>
          <string-name>
            <surname>Z.</surname>
          </string-name>
          , (
          <year>2004</year>
          ),
          <article-title>"Categorical data visualization and clustering using subjective factors", Data &amp; Knowledge Engineering, Published by Elsevier B</article-title>
          .V.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]. Chen N. and
          <string-name>
            <surname>Marques</surname>
            <given-names>N. C.</given-names>
          </string-name>
          , (
          <year>2005</year>
          ),
          <article-title>“An Extension of SelfOrganizing Maps to Categorical Data”</article-title>
          ,
          <source>Proceedings of the 12th</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]. Kaski
          <string-name>
            <surname>S.</surname>
          </string-name>
          , (
          <year>1997</year>
          ),
          <article-title>"Data exploration using self-organizing maps”</article-title>
          ,
          <source>Acta Polytechnica Scandinavica</source>
          , Mathematics, Computing and Management in Engineering Series No.
          <volume>82</volume>
          ,
          <year>Espoo 1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]. Kohonen
          <string-name>
            <surname>T</surname>
          </string-name>
          , (
          <year>1999</year>
          ), “
          <article-title>The Self-Organizing Map (SOM)”</article-title>
          , Helsinki University of Technology, Laboratory of Computer and Information Science,
          <source>Neural Networks Research Centre, Quinquennial Report (1994-1998)</source>
          ,
          <article-title>(Downloaded from http://www</article-title>
          .cis.hut.fi/research/reports/quinquennial/ January
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5].
          <source>Nigeria Climate Review</source>
          ,
          <year>2010</year>
          , Nigerian Meteorological Agency, www.nimetng.org
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]. Pampalk
          <string-name>
            <given-names>E</given-names>
            ,
            <surname>Rauber</surname>
          </string-name>
          <string-name>
            <given-names>A</given-names>
            ,
            <surname>Merkl</surname>
          </string-name>
          <string-name>
            <surname>D</surname>
          </string-name>
          , (
          <year>2002</year>
          ), “
          <article-title>Using Smoothed Data Histograms for Cluster Visualization in Self Organizing Maps”</article-title>
          ,
          <source>Technical Report OeFAI-TR-2002-29</source>
          , extended version published in
          <source>Proceedings of the International Conference on Artificial Neural Networks, Springer Lecture Notes in Computer Science</source>
          , Madrid, Spain,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>. Pelczer I. J.</given-names>
            and
            <surname>Cisneros</surname>
          </string-name>
          <string-name>
            <surname>H. L.</surname>
          </string-name>
          , (
          <year>2008</year>
          ), “
          <article-title>Identification of rainfall patterns over the Valley of Mexico”</article-title>
          , 11th International Conference on Urban Drainage, Edinburgh, Scotland, UK,
          <year>2008</year>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]. Principe
          <string-name>
            <given-names>J. C.</given-names>
            ,
            <surname>Euliano</surname>
          </string-name>
          <string-name>
            <surname>N. R. Lefebvre W. C</surname>
          </string-name>
          , (
          <year>2000</year>
          ),
          <source>Neural and Adaptive Systems: Fundamentals Through Simulations, John Wiley and Sons Inc, ISBN 0-471-35167-9</source>
          , pp
          <fpage>656</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]. Statsoft Electronic Statistics Textbook, (
          <year>2002</year>
          ), Copyright,
          <fpage>1984</fpage>
          -
          <lpage>2003</lpage>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]. Ultsch,
          <string-name>
            <surname>A.</surname>
          </string-name>
          , (
          <year>1999</year>
          ),
          <article-title>Data Mining and Knowledge Discovery with Emergent Self-Organizing Feature Maps for Multivariate Time Series</article-title>
          , In Kohonen Maps, (
          <year>1999</year>
          ), pp.
          <fpage>33</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]. Ultsch
          <string-name>
            <surname>A</surname>
          </string-name>
          , (
          <year>2003a</year>
          ),
          <article-title>Maps for the Visualization of highdimensional Data Spaces</article-title>
          ,
          <source>Proc. Workshop on Self Organizing Maps</source>
          , pp
          <fpage>225</fpage>
          -
          <lpage>230</lpage>
          , Kyushu, Japan,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]. Ultsch
          <string-name>
            <surname>A</surname>
          </string-name>
          , (
          <year>2003b</year>
          ), U*
          <article-title>-Matrix: a Tool to visualize Clusters in high dimensional Data</article-title>
          ,
          <source>Technical Report No. 36</source>
          , Computer Science Department, University of Marburg, Germany,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]. Ultsch
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Moerchen</surname>
          </string-name>
          <string-name>
            <surname>F</surname>
          </string-name>
          , (
          <year>2005</year>
          ),
          <article-title>ESOM-Maps: tools for clustering, visualization, and classification with Emergent SOM</article-title>
          ,
          <source>Technical Report No. 46</source>
          , Dept. of Mathematics and Computer Science, University of Marburg, Germany,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>. Wehrens R. Buydens L. M. C.</surname>
          </string-name>
          ,
          <year>2007</year>
          , Self- and
          <string-name>
            <surname>Superorganizing Maps</surname>
            in
            <given-names>R</given-names>
          </string-name>
          :
          <article-title>The kohonen Package, Journal of Statistical Software,published by the American Statistical Association</article-title>
          , Vol.
          <volume>21</volume>
          ,
          <string-name>
            <surname>Issue</surname>
            <given-names>5</given-names>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]. Zengyou He,
          <string-name>
            <surname>Xiaofe I Fe</surname>
          </string-name>
          , Shengchun Deng, (
          <year>2003</year>
          ), “
          <article-title>Clustering Mixed Categorical and Numeric Data”</article-title>
          , Department of Computer Science and Engineering, Harbin Institute of Technology, Harbin 150001,
          <string-name>
            <surname>P.</surname>
          </string-name>
          R. China
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>