<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Nonlinear Dimensionality Reduction Using Combined Approach to Feature Space Decomposition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Myasnikov E.V.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Samara State Aerospace University, Image Processing Systems Institute of the Russian Academy of Science Samara</institution>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>85</fpage>
      <lpage>95</lpage>
      <abstract>
        <p>In this paper we propose a new combined approach to feature space decomposition to improve the efficiency of the nonlinear dimensionality reduction method. The approach performs the decomposition of the original multidimensional space, taking into account the configuration of objects in the target low-dimensional space. The proposed approach is compared to the approach using hierarchical clustering in the original space and to the approach based on the decomposition of the target space using KD-Tree.</p>
      </abstract>
      <kwd-group>
        <kwd>dimensionality reduction</kwd>
        <kwd>multidimensional scaling</kwd>
        <kwd>Sammon's mapping</kwd>
        <kwd>hierarchical clustering</kwd>
        <kwd>KD-Tree</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Dimensionality reduction methods, operating on the principle of preserving the
pairwise distances between objects can be used as a means to display multidimensional
data in scientific research and in production activities in a number of areas: biology,
genetics, sociology, economics, etc. In modern information systems, such methods
can be used to create navigation systems for multimedia databases, as well as in
interface design to virtual directories. In the field of image analysis and processing
nonlinear dimensionality reduction techniques have been applied not only for research but
also to solve a number of applied problems: creation of automated systems for image
segmentation, thematic mapping from satellite imagery and others.</p>
      <p>One of the most well-known dimensionality reduction methods is a method [9],
which minimizes the following error of multidimensional data representation
 
d (oi , o j )  d *(oi , o j )2
d (oi , o j )</p>
      <p>
        (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
where oi is an object from a set O, d (oi ,o j ) is the distance between objects in the
original space, d *(oi ,o j ) is the distance between objects in the target space.
      </p>
      <p>
        Iterative procedure that minimizes the error (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) may be based on the following
recurrence relations for the coordinates yi in the target low-dimensional space (gradient
descent):
yi (t  1)  yi (t)  m  ς(oi , o j )
o j O
where (m can be calculated once):
m 
      </p>
      <p>
        2 
 d (oi , o j )
oi ,o j O
i j
ς(oi ,o j ) 
d (oi ,o j )  d *(oi ,o j )
d (oi ,o j )  d *(oi ,o j )  yi  y j 
It is worth noting that the performance of the nonlinear mapping has been improved
by extending multidimensional data representation error (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) with Bregman
divergences [11, 12].
      </p>
      <p>Unfortunately, a significant drawback of the nonlinear dimensionality reduction
techniques, operating according to the iteration scheme (similar to that shown above)
is the high computational complexity (O(K2) per iteration, where K  O is the
number of objects). For this reason a number of techniques with a reduced computational
complexity have been proposed [3, 4, 5, 6, 7].</p>
      <p>
        One of the most effective ways to reduce the computational complexity is the
hierarchical decomposition of space. After performing such decomposition objects can be
considered not individually, but in groups, that allows to speed up the iterative
optimization process. So, if all the objects of the original set O are divided into groups
s j  S , then (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) takes the form
yi (t 1)  yi (t)  m   ~ς(oi , s j )
      </p>
      <p>s j S
~ς(oi , s j )  s j  d (oi , s j )  d *(oi , s j )  yi  ys j </p>
      <p>
        d (oi , s j )  d *(oi , s j )
where d (oi , s j ) is the distance from the object to the center of the group sj in the
original space, d *(oi ,c j ) is the distance from the object to the center of the group in
the target space, y s j is the coordinates of the center of the group in the target space.
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
Such decomposition allows to reduce the computational complexity to O(LK) per
iteration, where L  S is the number of groups.
      </p>
      <p>
        It is obvious that the expression (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) allows to approximate (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) with a certain
accuracy that depends on how well objects are divided into groups. For this reason, there
is a question about how to better carry out such decomposition. It is understood that
the decomposition can be performed in both the source and target space. Approach
based on decomposition of the original space, is implemented in a nonlinear method
for dimensionality reduction of data using reference nodes [5] proposed by the author
of this article. Approach based on decomposition of the target space has been
implemented, for example, in [1] (regular decomposition) and [8] (hierarchical
decomposition) to solve the problem of drawing graphs to approximate the forces acting on the
vertices of the graph.
      </p>
      <p>In this paper we propose a new method that uses the combined approach to feature
space decomposition to improve the efficiency of the nonlinear dimensionality
reduction method. The proposed method is based on the nonlinear method for
dimensionality reduction of data using reference nodes [5] that performs the decomposition of
the original space, and complemented with the additional control on the configuration
of objects in the target space. The proposed method is compared to the method based
on the decomposition of the original space, and to the method based on the
decomposition of the target space, both in terms of the quality of the mapping, as well as in
terms of the operating time.</p>
      <p>The paper is organized as follows. The next section is devoted to the description of
the known approaches used in the present study, and consists of two subsections. The
third section is devoted to the description of the proposed method. The fourth section
presents the results of numerical experiments. At the end of the paper the conclusion
is given.
2
2.1</p>
      <p>Brief Description of the Methods Used in the Study
Nonlinear Method for Dimensionality Reduction of Data Using Reference
Nodes
Approach based on the decomposition of the original space, is implemented in a
nonlinear method for dimensionality reduction of data using reference nodes [5]. This
method consists of four steps, briefly described below.</p>
      <p>The inputs to the method are the feature vectors describing the objects in the
original multidimensional space. The outputs of the method are the feature vectors
describing the objects in the target low-dimensional space.</p>
      <p>Step 1: Construction of a hierarchy of clusters. At the first stage of the method
we make a hierarchical partition of the initial set of objects into clusters (subsets of
objects with similar characteristics in the original feature space). When we refer to the
hierarchy of clusters we mean tree-like data structure, the root of which is a cluster of
top-level, and each vertex-cluster contains either subclusters or the objects of the
original set.</p>
      <p>Step 2: Initialization of the coordinates of objects in the target space.
Initialization of the coordinates of objects in the target space can be performed in various
ways, e.g., using random values, the results of the PCA method, etc.</p>
      <p>Step 3: Construction of the lists of reference nodes. At the third stage, for each
object of the original set the lists of reference nodes are built. The reference node of
the object o refers to a certain object oi  o or group of objects oi i1..N that
possesses close characteristics in the original multidimensional space and considered as a
single object.</p>
      <p>Let o be an object for which we need to form a list S of the reference nodes and
let C be an arbitrary cluster of the hierarchy. By using the hierarchy of clusters
obtained at the first stage the construction of the reference nodes list may be
implemented using the following recursive algorithm.
1. If the cluster C satisfies the decomposition criterion with respect to the considered
object and subclusters Ci C, i  1..N are contained in cluster C , then one must
apply this recursive algorithm to each of the subclusters C1,C2 ,.., CN .
2. If the cluster C satisfies the decomposition criterion with respect to the considered
object and objects oi C,i  1..N are contained in cluster C , then one must:
(a) add all objects that close to the considered object to the list of reference nodes</p>
      <p>S ;
(b) distinguish the set V of objects situated at a distance from the considered
object as an incomplete cluster;
(c) in the case when this set V is nonempty, add it to the list of reference nodes.
3. If the cluster C doesn’t satisfy the decomposition criterion with respect to the
considered object then add it to the list of reference nodes S .</p>
      <p>Decomposition criterion used in this paper is slightly different from the criterion used
in [5] and restricts (by the value ) the estimation of the angle at which the
hypersphere of the cluster is observed from the object o in the original space.</p>
      <p>Figure 1 illustrates this process. The cluster, which is observed from an object o at
an angle  will be divided into three clusters if    . The cluster, which is
observed at an angle  will be considered as a whole if    .</p>
      <p>
        Step 4: Iterative optimization procedure. At the final stage of the method
iterative procedure is performed, which allows to clarify the position of objects in the
target low-dimensional space. The work of the iterative procedure is based on (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ), (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ).
2.2
      </p>
      <p>Nonlinear Dimensionality Reduction Method Using KD-Tree for the
target space decomposition
In contrast to the approach discussed above decomposition of the target space
inevitably leads to the need for periodic renewal or adjustment of the structure by which the
decomposition is performed. The reason for this consists in changing the coordinates
of objects in the target space as you perform the optimization process.</p>
      <p>In this paper, KD-trees are used for the decomposition of the target space. Tree
construction is performed at each iteration of the optimization process using the
following recursive procedure.
1. Calculation of the characteristics of the current tree node (average coordinates in
the source and target space, boundaries of the node in the target space).
2. In the case when the node contains only one object, then exit.
3. Splitting the node into the two child nodes using the perpendicular to the most
elongated boundary so that the number of objects in the child nodes is
approximately the same.
4. Perform this procedure for the newly formed child nodes.</p>
      <p>
        Constructed tree is used to perform the next iteration of the optimization process. For
each object, the sum (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) is calculated by traversing the tree as follows.
1. If the decomposition criterion is not satisfied and the objects contained in the
current node of the tree can be considered as a group, then the next term in (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) is
calculated using (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ).
2. Otherwise decomposition criterion is satisfied, and both subtrees (child nodes)
traversed similarly.
A decomposition criterion used in the tree traversal is similar to the above, and
restricts (by the value ) the estimation of the angle at which the minimum bounding
rectangle of the node is observed from the position of the object o in the target space.
      </p>
      <p>Figure 2 illustrates this process. Different decomposition levels are shown using
different lines: the first level is shown by the solid line, the second level is shown by
the dash-dot line, the third level is shown by the dashed line. A group of four objects
(minimum bounding rectangle is shown by dots) will be devided into two groups if
the angle  at which the minimum bounding rectangle is observed from the position
of the object o is greater than the decomposition angle  .
It is worth noting that this algorithm is executed at each iteration for each object, so to
improve the efficiency the constructed tree is traversed not recursively, but iteratively
using special pointers.</p>
      <p>Note that some additional experimental results on comparison of methods
described in this section can be found in [6].
3</p>
    </sec>
    <sec id="sec-2">
      <title>Proposed Method</title>
      <p>
        As noted above the expression (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) approximates (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) with some accuracy. The error
of approximation for some object o and the set s of individual objects can be
represented using (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) as follows:
 (o, s)   d (o, o j )  d *(o, o j )  y  y j ~ς(o, s) .
      </p>
      <p>
        o j s d (o, o j )  d *(o, o j )
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
      </p>
      <p>Unfortunately, the use of the described above approaches can cause significant (in
some cases unbounded) approximation errors as in the first case the decomposition
criterion takes into account only the characteristics of nodes in the original space and
in the second case criterion takes into account only the characteristics of nodes in the
target space. The way out of this situation may be a combination of the original space
decomposition with the additional control on the configuration of objects in the target
space.</p>
      <p>In this paper, for this purpose a nonlinear method for dimensionality reduction of
data using reference nodes is complemented by the preservation and analysis of the
boundaries of nodes in the target space. The proposed method consists of the same
steps as the base method [5]:
1. Construction of a hierarchy of clusters
2. Initialization of objects coordinates in the target space
3. Construction of reference nodes lists
4. Iterative optimization procedure</p>
      <p>At the first step the difference between the proposed method and the base one is in
the structure of nodes in the cluster tree. In the proposed method each node contains
information about its boundaries in the target low-dimensional space. The second and
the third steps are the same as in the base method. At the fourth step iterative
optimization procedure is different in the proposed method.</p>
      <p>
        At the each step of the optimization process for each object oi O with the
corresponding reference nodes list Si it recalculates the coordinates yi in the target
lowdimensional space. To do this according to (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) it iterates through the list of reference
nodes Si and calculates the approximated term for a given node s j  Si using the
following recursive algorithm.
1. If the given node s j is the object oj O of the initial set of objects, then the
exact value of the corresponding term is calculated using (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ).
2. If the decomposition criterion is not satisfied for given node s j in the target
lowdimensional space and the objects contained in the given node can be considered as
a group, then the approximated term is calculated using (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ).
3. Otherwise the decomposition criterion is satisfied for the given node s j in the
target low-dimensional space, and all subtrees of the given node in the cluster tree
traversed similarly using this algorithm.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Experimental Study</title>
      <p>In this paper all the studied methods have been implemented in C++ language in the
integrated development environment Borland Turbo C++ 2006 Explorer. The PC
based on Intel Core i5-3470 CPU 3.2 GHz has been used in the experiments. Corel
Image Features Data Set (http://archive.ics.uci.edu/ml/databases/CorelFeatures/
CorelFeatures.data.html) has been used in the experiments as an initial dataset. This
dataset contains a set of features, calculated from the digital images from the Corel
image collection (http://corel.digitalriver.com/). In particular, the following feature
sets have been used.</p>
      <p>Feature set 1 contains color moments [10]. Three features were calculated for each
color component: mean, standard deviation, and skewness. The dimensionality of the
feature space is 9.</p>
      <p>Feature set 2 contains color histograms [13] constructed in the HSV color space.
Color space was divided into 8 ranges of H and 4 ranges of S. The dimensionality of
the feature space is 32.</p>
      <p>Feature set 3 contains texture features based on co-occurrence matrices [2]. Four
co-occurrence features (second angular moment, contrast, inverse difference moment,
and entropy) were computed in four directions (horizontal, vertical, and two
diagonal). The dimensionality of the feature space is 16.</p>
      <p>Fragments containing features information for the required number of images has
been selected from these feature sets. Then for sets 1 and 3, the identity component
has been added to the feature information and normalization has been performed.</p>
      <p>
        To evaluate effectiveness of the methods the value of multidimensional data
representation error (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) has been calculated and the average per iteration execution time of
the optimization procedure has been measured.
      </p>
      <p>Work of the methods stopped when the rate of decreasing the error slowed down
(i.e. the relative decrease in error for ten iterations does not exceed 0.05). In all cases,
the dimension of the target space has been set equal to two (two-dimensional data
mapping). Initialization is performed using the PCA method.</p>
      <p>Some results for the feature set 1 are shown in Fig. 3-6.</p>
      <p>Fig. 3 and 4 show the dependence of the qualitative and temporal characteristics on
the decomposition angle  at which the algorithm moves to the child nodes of the
corresponding hierarchical structure (cluster tree or KD-tree).</p>
      <p>As can be seen from these results, the large values of the angle  leads to the
expected deterioration of the mapping quality due to a coarser approximation, which is
reflected in the higher values of the multidimensional data representation error  (see
Fig. 3). The time it takes to perform a single iteration, decreases with increasing angle
 (see Fig. 4), due to the smaller number of processed nodes of the corresponding
hierarchical structure during the construction of approximations.</p>
      <p>It should be noted that the approach using the original space decomposition
provides less value of multidimensional data representation error  than the approach
using the decomposition of the target space.
Fig. 3. Dependence of the multidimensional data representation error </p>
      <p>on the decomposition angle  (in fractions of  radian)
Fig. 4. Dependence of the average execution time per iteration (in ms)</p>
      <p>on the decomposition angle  (in fractions of  radian)
At the same time the average execution time per iteration is much smaller when using
the approach with the decomposition of the original space in the range
[0.05 , 0.2 ].</p>
      <p>The use of the combined approach allows slightly reduce the multidimensional data
representation error in comparison with the approach based on the decomposition of
the original space. At the same time the average time per iteration using a combined
approach is significantly larger than when using other considered approaches.</p>
      <p>Dependencies of quality and temporal characteristics on the number of objects
shown in the figures 5 and 6 confirm said above. The approach using the original
space decomposition provides less value of multidimensional data representation error
 than the approach using the decomposition of the target space (fig. 5a). The quality
of mapping, measured by the multidimensional data representation error , formed
using decomposition in the original feature space is only slightly inferior to the
combined approach (fig. 5b). The quality of mapping formed using combined approach is
virtually identical to the quality obtained with the base method without
decomposition.</p>
      <p>At the same time approach, using the decomposition of the original space, allows
us to construct mapping much faster than the other considered approaches (Fig. 6).
Fig. 5. Dependence of the multidimensional data representation error </p>
      <p>on the number of objects</p>
      <p>Note that the experiments performed on other datasets described above, show
similar results.
In this paper, we conducted a study of different approaches to the decomposition of
space in terms of quality and speed of the nonlinear dimensionality reduction method.
The study showed that for the presented data sets decomposition of the original space
is more preferable than the decomposition of target space in terms of both quality and
time. Combining the decomposition of the original space with the decomposition of
target space has allowed slightly improve quality.</p>
      <p>Acknowledgments. This work was financially supported by the Ministry of
education and science of the Russian Federation in the framework of the implementation of
the Program of increasing the competitiveness of SSAU among the world's leading
scientific and educational centers for 2013-2020 years and by the Russian Foundation
for Basic Research, project № 15-07-01164 -a.
6</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Fruchterman</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reingold</surname>
          </string-name>
          , E.:
          <article-title>Graph Drawing by Force-directed Placement</article-title>
          .
          <source>Software - Practice and Experience</source>
          , vol.
          <volume>21</volume>
          , no.
          <issue>11</issue>
          , pp.
          <fpage>1129</fpage>
          -
          <lpage>1164</lpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Haralick</surname>
            ,
            <given-names>R.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shanmugam</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dinstein</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Texture Features for Image Classification</article-title>
          .
          <source>IEEE Trans. on Sys. Man. and Cyb</source>
          . SMC-
          <volume>3</volume>
          (
          <issue>6</issue>
          ) (
          <year>1973</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>R.C.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Slagle</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blum</surname>
            ,
            <given-names>H. A.</given-names>
          </string-name>
          :
          <article-title>Triangulation Method for the Sequential Mapping of Points from N-Space to Two-Space</article-title>
          .
          <source>IEEE Transactions on Computers</source>
          , vol.
          <volume>26</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>288</fpage>
          -
          <lpage>292</lpage>
          (
          <year>1977</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Morrison</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ross</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chalmers</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Fast Multidimensional Scaling through Sampling, Springs and Interpolation</article-title>
          .
          <source>Information Visualization</source>
          , vol.
          <volume>2</volume>
          , pp.
          <fpage>68</fpage>
          -
          <lpage>77</lpage>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Myasnikov</surname>
            ,
            <given-names>E.V.</given-names>
          </string-name>
          :
          <article-title>A Nonlinear Method for Dimensionality Reduction of Data Using Reference Nodes. Pattern Recognition and Image Analysis</article-title>
          , vol.
          <volume>22</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>337</fpage>
          -
          <lpage>345</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Myasnikov</surname>
            ,
            <given-names>E.V.</given-names>
          </string-name>
          :
          <article-title>The Choice of a Method for Feature Space Decomposition for NonLinear Dimensionality Reduction</article-title>
          .
          <source>Computer optics</source>
          , vol.
          <volume>38</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>790</fpage>
          -
          <lpage>797</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Pekalska</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>de Ridder</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duin</surname>
            ,
            <given-names>R.P.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kraaijveld</surname>
            ,
            <given-names>M.A.:</given-names>
          </string-name>
          <article-title>A New Method of Generalizing Sammon Mapping with Application to Algorithm Speed-up</article-title>
          .
          <source>Proc. ASCI'99, 5th Annual Conf. of the Advanced School for Computing and Imaging</source>
          , pp.
          <fpage>221</fpage>
          -
          <lpage>228</lpage>
          . Heijen,
          <string-name>
            <surname>Netherlands</surname>
          </string-name>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Quigley</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eades</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : FADE: Graph Drawing, Clustering, and
          <string-name>
            <given-names>Visual</given-names>
            <surname>Abstraction</surname>
          </string-name>
          .
          <source>Proceedings of the 8th International Symposium on Graph Drawing</source>
          , pp.
          <fpage>197</fpage>
          -
          <lpage>210</lpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sammon</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          <string-name>
            <surname>Jr</surname>
          </string-name>
          .
          <article-title>: A Nonlinear Mapping for Data Structure Analysis</article-title>
          .
          <source>IEEE Transactions on Computers</source>
          , vol. C-
          <volume>18</volume>
          , no.
          <issue>5</issue>
          , pp.
          <fpage>401</fpage>
          -
          <lpage>409</lpage>
          (
          <year>1969</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Stricker</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Orengo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Similarity of color images</article-title>
          .
          <source>In Proc. SPIE Conf. on Vis. Commun. and Image Proc</source>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crowe</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fyfe</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Extending metric multidimensional scaling with Bregman divergences</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>44</volume>
          (
          <issue>5</issue>
          ), pp.
          <fpage>1137</fpage>
          -
          <lpage>1154</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fyfe</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crowe</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Extending Sammon mapping with Bregman divergences</article-title>
          .
          <source>Information Sciences</source>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Swain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ballard</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Color indexing</article-title>
          .
          <source>International Journal of Computer Vision</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ) (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>