=Paper=
{{Paper
|id=Vol-1452/paper11
|storemode=property
|title=A Nonlinear Dimensionality Reduction Using Combined Approach to Feature Space Decomposition
|pdfUrl=https://ceur-ws.org/Vol-1452/paper11.pdf
|volume=Vol-1452
|dblpUrl=https://dblp.org/rec/conf/aist/Myasnikov15
}}
==A Nonlinear Dimensionality Reduction Using Combined Approach to Feature Space Decomposition==
A Nonlinear Dimensionality Reduction Using Combined Approach to Feature Space Decomposition Myasnikov E.V. Samara State Aerospace University, Image Processing Systems Institute of the Russian Academy of Science Samara, Russia mevg@geosamara.ru Abstract. In this paper we propose a new combined approach to feature space decomposition to improve the efficiency of the nonlinear dimensionality reduc- tion method. The approach performs the decomposition of the original multidi- mensional space, taking into account the configuration of objects in the target low-dimensional space. The proposed approach is compared to the approach us- ing hierarchical clustering in the original space and to the approach based on the decomposition of the target space using KD-Tree. Keywords: dimensionality reduction; multidimensional scaling; Sammon's mapping; hierarchical clustering; KD-Tree 1 Introduction Dimensionality reduction methods, operating on the principle of preserving the pair- wise distances between objects can be used as a means to display multidimensional data in scientific research and in production activities in a number of areas: biology, genetics, sociology, economics, etc. In modern information systems, such methods can be used to create navigation systems for multimedia databases, as well as in inter- face design to virtual directories. In the field of image analysis and processing nonlin- ear dimensionality reduction techniques have been applied not only for research but also to solve a number of applied problems: creation of automated systems for image segmentation, thematic mapping from satellite imagery and others. One of the most well-known dimensionality reduction methods is a method [9], which minimizes the following error of multidimensional data representation d (o , o ) d (o , o ) * 2 1 i j i j (1) d (oi , o j ) d (o , o ) oi , o j O i j oi , o j O i j i j 85 where oi is an object from a set O, d (oi , o j ) is the distance between objects in the original space, d * (oi , o j ) is the distance between objects in the target space. Iterative procedure that minimizes the error (1) may be based on the following re- currence relations for the coordinates yi in the target low-dimensional space (gradient descent): y i (t 1) y i (t ) m ς(o , o ) o j O i j (2) where (m can be calculated once): 2 m d (o , o ) oi , o j O i j i j d (oi , o j ) d * (oi , o j ) ς(oi , o j ) yi y j (3) d (oi , o j ) d * (oi , o j ) It is worth noting that the performance of the nonlinear mapping has been improved by extending multidimensional data representation error (1) with Bregman divergenc- es [11, 12]. Unfortunately, a significant drawback of the nonlinear dimensionality reduction techniques, operating according to the iteration scheme (similar to that shown above) is the high computational complexity (O(K2) per iteration, where K O is the num- ber of objects). For this reason a number of techniques with a reduced computational complexity have been proposed [3, 4, 5, 6, 7]. One of the most effective ways to reduce the computational complexity is the hier- archical decomposition of space. After performing such decomposition objects can be considered not individually, but in groups, that allows to speed up the iterative opti- mization process. So, if all the objects of the original set O are divided into groups s j S , then (2) takes the form y i (t 1) y i (t ) m ~ς (o , s ) s j S i j (4) d (oi , s j ) d (oi , s j ) * ~ ς (oi , s j ) s j yi y s j (5) d (oi , s j ) d * (oi , s j ) where d (oi , s j ) is the distance from the object to the center of the group sj in the original space, d * (oi , c j ) is the distance from the object to the center of the group in the target space, y s j is the coordinates of the center of the group in the target space. 86 Such decomposition allows to reduce the computational complexity to O(LK) per iteration, where L S is the number of groups. It is obvious that the expression (4) allows to approximate (2) with a certain accu- racy that depends on how well objects are divided into groups. For this reason, there is a question about how to better carry out such decomposition. It is understood that the decomposition can be performed in both the source and target space. Approach based on decomposition of the original space, is implemented in a nonlinear method for dimensionality reduction of data using reference nodes [5] proposed by the author of this article. Approach based on decomposition of the target space has been imple- mented, for example, in [1] (regular decomposition) and [8] (hierarchical decomposi- tion) to solve the problem of drawing graphs to approximate the forces acting on the vertices of the graph. In this paper we propose a new method that uses the combined approach to feature space decomposition to improve the efficiency of the nonlinear dimensionality reduc- tion method. The proposed method is based on the nonlinear method for dimension- ality reduction of data using reference nodes [5] that performs the decomposition of the original space, and complemented with the additional control on the configuration of objects in the target space. The proposed method is compared to the method based on the decomposition of the original space, and to the method based on the decompo- sition of the target space, both in terms of the quality of the mapping, as well as in terms of the operating time. The paper is organized as follows. The next section is devoted to the description of the known approaches used in the present study, and consists of two subsections. The third section is devoted to the description of the proposed method. The fourth section presents the results of numerical experiments. At the end of the paper the conclusion is given. 2 Brief Description of the Methods Used in the Study 2.1 Nonlinear Method for Dimensionality Reduction of Data Using Reference Nodes Approach based on the decomposition of the original space, is implemented in a non- linear method for dimensionality reduction of data using reference nodes [5]. This method consists of four steps, briefly described below. The inputs to the method are the feature vectors describing the objects in the origi- nal multidimensional space. The outputs of the method are the feature vectors describ- ing the objects in the target low-dimensional space. Step 1: Construction of a hierarchy of clusters. At the first stage of the method we make a hierarchical partition of the initial set of objects into clusters (subsets of objects with similar characteristics in the original feature space). When we refer to the hierarchy of clusters we mean tree-like data structure, the root of which is a cluster of 87 top-level, and each vertex-cluster contains either subclusters or the objects of the orig- inal set. Step 2: Initialization of the coordinates of objects in the target space. Initializa- tion of the coordinates of objects in the target space can be performed in various ways, e.g., using random values, the results of the PCA method, etc. Step 3: Construction of the lists of reference nodes. At the third stage, for each object of the original set the lists of reference nodes are built. The reference node of the object o refers to a certain object oi o or group of objects oi i 1.. N that pos- sesses close characteristics in the original multidimensional space and considered as a single object. Let o be an object for which we need to form a list S of the reference nodes and let C be an arbitrary cluster of the hierarchy. By using the hierarchy of clusters ob- tained at the first stage the construction of the reference nodes list may be implement- ed using the following recursive algorithm. 1. If the cluster C satisfies the decomposition criterion with respect to the considered object and subclusters Ci C, i 1..N are contained in cluster C , then one must apply this recursive algorithm to each of the subclusters C1, C2 ,.., CN . 2. If the cluster C satisfies the decomposition criterion with respect to the considered object and objects oi C, i 1..N are contained in cluster C , then one must: (a) add all objects that close to the considered object to the list of reference nodes S; (b) distinguish the set V of objects situated at a distance from the considered ob- ject as an incomplete cluster; (c) in the case when this set V is nonempty, add it to the list of reference nodes. 3. If the cluster C doesn’t satisfy the decomposition criterion with respect to the con- sidered object then add it to the list of reference nodes S . Decomposition criterion used in this paper is slightly different from the criterion used in [5] and restricts (by the value ) the estimation of the angle at which the hypersphere of the cluster is observed from the object o in the original space. Figure 1 illustrates this process. The cluster, which is observed from an object o at an angle will be divided into three clusters if . The cluster, which is ob- served at an angle will be considered as a whole if . 88 Fig. 1. Illustration to the construction of a list of reference nodes Step 4: Iterative optimization procedure. At the final stage of the method iterative procedure is performed, which allows to clarify the position of objects in the target low-dimensional space. The work of the iterative procedure is based on (4), (5). 2.2 Nonlinear Dimensionality Reduction Method Using KD-Tree for the target space decomposition In contrast to the approach discussed above decomposition of the target space inevita- bly leads to the need for periodic renewal or adjustment of the structure by which the decomposition is performed. The reason for this consists in changing the coordinates of objects in the target space as you perform the optimization process. In this paper, KD-trees are used for the decomposition of the target space. Tree construction is performed at each iteration of the optimization process using the fol- lowing recursive procedure. 1. Calculation of the characteristics of the current tree node (average coordinates in the source and target space, boundaries of the node in the target space). 2. In the case when the node contains only one object, then exit. 3. Splitting the node into the two child nodes using the perpendicular to the most elongated boundary so that the number of objects in the child nodes is approxi- mately the same. 4. Perform this procedure for the newly formed child nodes. Constructed tree is used to perform the next iteration of the optimization process. For each object, the sum (4) is calculated by traversing the tree as follows. 1. If the decomposition criterion is not satisfied and the objects contained in the cur- rent node of the tree can be considered as a group, then the next term in (4) is cal- culated using (5). 2. Otherwise decomposition criterion is satisfied, and both subtrees (child nodes) traversed similarly. 89 A decomposition criterion used in the tree traversal is similar to the above, and re- stricts (by the value ) the estimation of the angle at which the minimum bounding rectangle of the node is observed from the position of the object o in the target space. Figure 2 illustrates this process. Different decomposition levels are shown using different lines: the first level is shown by the solid line, the second level is shown by the dash-dot line, the third level is shown by the dashed line. A group of four objects (minimum bounding rectangle is shown by dots) will be devided into two groups if the angle at which the minimum bounding rectangle is observed from the position of the object o is greater than the decomposition angle . Fig. 2. Illustration of decomposition in the target space using KD-Tree It is worth noting that this algorithm is executed at each iteration for each object, so to improve the efficiency the constructed tree is traversed not recursively, but iteratively using special pointers. Note that some additional experimental results on comparison of methods de- scribed in this section can be found in [6]. 3 Proposed Method As noted above the expression (4) approximates (2) with some accuracy. The error of approximation for some object o and the set s of individual objects can be repre- sented using (5) as follows: d (o, o j ) d * (o, o j ) (o, s) y y j ~ ς (o, s) . (6) o j s d (o, o j ) d (o, o j ) * 90 Unfortunately, the use of the described above approaches can cause significant (in some cases unbounded) approximation errors as in the first case the decomposition criterion takes into account only the characteristics of nodes in the original space and in the second case criterion takes into account only the characteristics of nodes in the target space. The way out of this situation may be a combination of the original space decomposition with the additional control on the configuration of objects in the target space. In this paper, for this purpose a nonlinear method for dimensionality reduction of data using reference nodes is complemented by the preservation and analysis of the boundaries of nodes in the target space. The proposed method consists of the same steps as the base method [5]: 1. Construction of a hierarchy of clusters 2. Initialization of objects coordinates in the target space 3. Construction of reference nodes lists 4. Iterative optimization procedure At the first step the difference between the proposed method and the base one is in the structure of nodes in the cluster tree. In the proposed method each node contains information about its boundaries in the target low-dimensional space. The second and the third steps are the same as in the base method. At the fourth step iterative optimi- zation procedure is different in the proposed method. At the each step of the optimization process for each object oi O with the corre- sponding reference nodes list S i it recalculates the coordinates yi in the target low- dimensional space. To do this according to (4) it iterates through the list of reference nodes S i and calculates the approximated term for a given node s j Si using the following recursive algorithm. 1. If the given node s j is the object o j O of the initial set of objects, then the ex- act value of the corresponding term is calculated using (3). 2. If the decomposition criterion is not satisfied for given node s j in the target low- dimensional space and the objects contained in the given node can be considered as a group, then the approximated term is calculated using (5). 3. Otherwise the decomposition criterion is satisfied for the given node s j in the tar- get low-dimensional space, and all subtrees of the given node in the cluster tree traversed similarly using this algorithm. 4 Experimental Study In this paper all the studied methods have been implemented in C++ language in the integrated development environment Borland Turbo C++ 2006 Explorer. The PC 91 based on Intel Core i5-3470 CPU 3.2 GHz has been used in the experiments. Corel Image Features Data Set (http://archive.ics.uci.edu/ml/databases/CorelFeatures/ CorelFeatures.data.html) has been used in the experiments as an initial dataset. This dataset contains a set of features, calculated from the digital images from the Corel image collection (http://corel.digitalriver.com/). In particular, the following feature sets have been used. Feature set 1 contains color moments [10]. Three features were calculated for each color component: mean, standard deviation, and skewness. The dimensionality of the feature space is 9. Feature set 2 contains color histograms [13] constructed in the HSV color space. Color space was divided into 8 ranges of H and 4 ranges of S. The dimensionality of the feature space is 32. Feature set 3 contains texture features based on co-occurrence matrices [2]. Four co-occurrence features (second angular moment, contrast, inverse difference moment, and entropy) were computed in four directions (horizontal, vertical, and two diago- nal). The dimensionality of the feature space is 16. Fragments containing features information for the required number of images has been selected from these feature sets. Then for sets 1 and 3, the identity component has been added to the feature information and normalization has been performed. To evaluate effectiveness of the methods the value of multidimensional data repre- sentation error (1) has been calculated and the average per iteration execution time of the optimization procedure has been measured. Work of the methods stopped when the rate of decreasing the error slowed down (i.e. the relative decrease in error for ten iterations does not exceed 0.05). In all cases, the dimension of the target space has been set equal to two (two-dimensional data mapping). Initialization is performed using the PCA method. Some results for the feature set 1 are shown in Fig. 3-6. Fig. 3 and 4 show the dependence of the qualitative and temporal characteristics on the decomposition angle at which the algorithm moves to the child nodes of the corresponding hierarchical structure (cluster tree or KD-tree). As can be seen from these results, the large values of the angle leads to the ex- pected deterioration of the mapping quality due to a coarser approximation, which is reflected in the higher values of the multidimensional data representation error (see Fig. 3). The time it takes to perform a single iteration, decreases with increasing angle (see Fig. 4), due to the smaller number of processed nodes of the corresponding hierarchical structure during the construction of approximations. It should be noted that the approach using the original space decomposition pro- vides less value of multidimensional data representation error than the approach using the decomposition of the target space. 92 (a) (b) Fig. 3. Dependence of the multidimensional data representation error on the decomposition angle (in fractions of radian) Fig. 4. Dependence of the average execution time per iteration (in ms) on the decomposition angle (in fractions of radian) At the same time the average execution time per iteration is much smaller when using the approach with the decomposition of the original space in the range [0.05 , 0.2 ]. The use of the combined approach allows slightly reduce the multidimensional data representation error in comparison with the approach based on the decomposition of the original space. At the same time the average time per iteration using a combined approach is significantly larger than when using other considered approaches. Dependencies of quality and temporal characteristics on the number of objects shown in the figures 5 and 6 confirm said above. The approach using the original space decomposition provides less value of multidimensional data representation error than the approach using the decomposition of the target space (fig. 5a). The quality of mapping, measured by the multidimensional data representation error , formed using decomposition in the original feature space is only slightly inferior to the com- bined approach (fig. 5b). The quality of mapping formed using combined approach is virtually identical to the quality obtained with the base method without decomposi- tion. At the same time approach, using the decomposition of the original space, allows us to construct mapping much faster than the other considered approaches (Fig. 6). 93 Figure 5a also shows the multidimensional data representation error obtained after initialization of objects coordinates in the target low-dimensional space using the PCA. (a) (b) Fig. 5. Dependence of the multidimensional data representation error on the number of objects Note that the experiments performed on other datasets described above, show simi- lar results. Fig. 6. Dependence of the average execution time per iteration (in ms) on the number of objects 5 Conclusion In this paper, we conducted a study of different approaches to the decomposition of space in terms of quality and speed of the nonlinear dimensionality reduction method. The study showed that for the presented data sets decomposition of the original space is more preferable than the decomposition of target space in terms of both quality and time. Combining the decomposition of the original space with the decomposition of target space has allowed slightly improve quality. 94 Acknowledgments. This work was financially supported by the Ministry of educa- tion and science of the Russian Federation in the framework of the implementation of the Program of increasing the competitiveness of SSAU among the world's leading scientific and educational centers for 2013-2020 years and by the Russian Foundation for Basic Research, project № 15-07-01164 -a. 6 References 1. Fruchterman, T., Reingold, E.: Graph Drawing by Force-directed Placement. Software – Practice and Experience, vol. 21, no. 11, pp. 1129-1164 (1991) 2. Haralick, R.M., Shanmugam, K., Dinstein, I.: Texture Features for Image Classification. IEEE Trans. on Sys. Man. and Cyb. SMC-3(6) (1973) 3. Lee, R.C.T., Slagle, J.R., Blum, H. A.: Triangulation Method for the Sequential Mapping of Points from N-Space to Two-Space. IEEE Transactions on Computers, vol. 26, no. 3, pp. 288-292 (1977) 4. Morrison, A., Ross, G., Chalmers, M.: Fast Multidimensional Scaling through Sampling, Springs and Interpolation. Information Visualization, vol. 2, pp.68-77 (2003) 5. Myasnikov, E.V.: A Nonlinear Method for Dimensionality Reduction of Data Using Ref- erence Nodes. Pattern Recognition and Image Analysis, vol. 22, no. 2, pp. 337–345 (2012) 6. Myasnikov, E.V.: The Choice of a Method for Feature Space Decomposition for Non- Linear Dimensionality Reduction. Computer optics, vol. 38, no. 4, pp. 790-797 (2014) 7. Pekalska, E., de Ridder, D., Duin, R.P.W., Kraaijveld, M.A.: A New Method of General- izing Sammon Mapping with Application to Algorithm Speed-up. Proc. ASCI'99, 5th An- nual Conf. of the Advanced School for Computing and Imaging, pp. 221-228. Heijen, Netherlands (1999) 8. Quigley, A., Eades, P.: FADE: Graph Drawing, Clustering, and Visual Abstraction. Pro- ceedings of the 8th International Symposium on Graph Drawing, pp. 197–210 (2001) 9. Sammon, J.W. Jr.: A Nonlinear Mapping for Data Structure Analysis. IEEE Transactions on Computers, vol. C-18, no. 5, pp. 401-409 (1969) 10. Stricker, M., Orengo, M.: Similarity of color images. In Proc. SPIE Conf. on Vis. Commun. and Image Proc (1995) 11. Sun, J., Crowe, M., Fyfe, C.: Extending metric multidimensional scaling with Bregman di- vergences. Pattern Recognition, 44 (5), pp. 1137–1154 (2011) 12. Sun, J., Fyfe, C., Crowe, M.: Extending Sammon mapping with Bregman divergences. In- formation Sciences (2011) 13. Swain, M., Ballard, D.: Color indexing. International Journal of Computer Vision, 7(1) (1991) 95