Research Article | Open Access

Fuding Xie, Yutao Fan, Ming Zhou, "Dimensionality Reduction by Weighted Connections between Neighborhoods", *Abstract and Applied Analysis*, vol. 2014, Article ID 928136, 5 pages, 2014. https://doi.org/10.1155/2014/928136

# Dimensionality Reduction by Weighted Connections between Neighborhoods

**Academic Editor:**Caihong Li

#### Abstract

Dimensionality reduction is the transformation of high-dimensional data into a meaningful representation of reduced dimensionality. This paper introduces a dimensionality reduction technique by weighted connections between neighborhoods to improve -Isomap method, attempting to preserve perfectly the relationships between neighborhoods in the process of dimensionality reduction. The validity of the proposal is tested by three typical examples which are widely employed in the algorithms based on manifold. The experimental results show that the local topology nature of dataset is preserved well while transforming dataset in high-dimensional space into a new dataset in low-dimensionality by the proposed method.

#### 1. Introduction

Real-world data, such as speech signals, digital photographs, or fMRI scans, usually has a high dimensionality. In order to handle such real-world data adequately, its dimensionality needs to be reduced. Ideally, the reduced representation should have a dimensionality that corresponds to the intrinsic dimensionality of the data. Fortunately, in many of those applications, all of the components of those high-dimensional data vectors are not independent of each other and in many cases the data points can be considered as lying on or close to a low-dimensional nonlinear manifold embedded in a high-dimensional space. Dimensionality reduction methods can also be thought of as a principled way to understand the high-dimensional data.

Mathematically, the problem of dimensionality reduction can be described as follows briefly. Assume we have a dataset with samples in -dimensional space. Assume further that this dataset has intrinsic dimensionality (where and often ). The intrinsic dimensionality means that the points in dataset are lying on or near a manifold with dimensionality that is embedded in the -dimensional space. The manifold may be non-Riemannian because of discontinuities. Dimensionality reduction techniques transform dataset with dimensionality into a new dataset with dimensionality , while retaining the geometry of the data as much as possible [1].

A large number of nonlinear techniques for dimensionality reduction have been proposed in the past decades. Among them, principal component analysis (PCA) [2], linear discriminant analysis (LDA) [3], and multidimensional scaling (MDS) [4] are three most classical techniques for dimensionality reduction. PCA finds a low-dimensional embedding of the data points that best preserves their variance as measured in the high-dimensional input space. Classical MDS finds an embedding that preserves the interpoint distances, equivalent to PCA when those distances are Euclidean. LDA searches the projection axes on which the not-same-class points are far from each other while requiring the same-class points to be close to each other. In 2000, two novel methods for manifold learning, isometric feature map method (Isomap) [5] and the locally linear embedding method (LLE) [6], have drawn great interests. Unlike other nonlinear dimension reduction methods, both LLE and Isomap methods emphasize simple algorithmic implementation and avoid nonlinear optimization formulations that are prone to local minima [7]. In Isomap, the geodesic distances between the data points are computed by constructing a neighborhood graph , in which every data point is connected with its nearest neighbors. The geodesic distances between all data points in dataset are computed, thereby forming a pairwise geodesic distance matrix. The low-dimensional representations are obtained by applying classical MDS on the resulting pairwise geodesic distance matrix. In LLE, the local properties of the data manifold are constructed by writing the high-dimensional data points as a linear combination of their nearest neighbors. In the low-dimensional representation of the data, LLE attempts to retain the reconstruction weights in the linear combinations as good as possible. van der Maaten et al. [1] presented a review and systematic comparison of a variety of nonlinear dimensionality reduction techniques which include Kernel PCA, maximum variance unfolding, diffusion maps, Laplacian eigenmaps, local tangent space analysis, Sammon mapping, multilayer autoencoders, locally linear coordination, manifold charting, Isomap, and LLE. The performances of these nonlinear techniques are investigated on artificial and natural tasks.

Recently, a great many methods of dimensionality reduction methods based on manifold learning have been introduced and some classical techniques have been further improved [8–15]. In [8], Hu et al. proposed a new dimensionality reduction algorithm called discriminant multidimensional mapping (DMM), which combines the advantages of multidimensional scaling (MDS) and LDA. DMM is effective for small sample datasets with high dimensionality. A dimensionality reduction technique, named orthogonal isometric projection (OIP), is proposed in [9]. In contrast with Isomap, which learns the low-dimension embedding and solves problem under the classic multidimensional scaling (MDS) framework, they consider an explicit linear projection by capturing the geodesic distance, which is able to handle new data straightforwardly and leads to a standard eigenvalue problem. To address the problem of finding a template function that represents the common pattern of a sample of curves, a novel algorithm based on a robust version of the isometric featuring mapping (Isomap) algorithm is developed by Dimeglio et al. [11]. H. Choi and S. Choi [12] presented a robust kernel Isomap which relates the Isomap to Mercer kernel machines, so that the generalization property naturally emerges, through kernel principal component analysis. However, it is still a challenging problem to further propose a novel method to reduce the dataset in high-dimensional space into a desired dataset in low-dimensionality; meanwhile, the local topology nature of dataset is preserved perfectly.

Motivated by the Isomap and LLE dimensionality reduction methods proposed by Tenenbaum et al. [5] and Roweis and Saul [6], respectively, we in this paper improve the -Isomap algorithm by local weighted connection between neighbors. The nearness between neighbors to a data point can be reflected well by the proposed technique. The desired dimensionality reduction results have been obtained on the three classical examples by the proposal.

The paper is organized as follows. Section 2 introduces classical MDS and Isomap dimensionality reduction techniques. An improvement of -Isomap algorithm is depicted in Section 3. Section 4 shows the validity of the proposal by two well-known datasets which are widely used by a large number of dimensionality reduction techniques. The conclusion is given in the last section.

#### 2. MDS and Isomap Techniques

##### 2.1. Classical MDS

MDS [4, 8, 16] is one of the global nonlinear techniques for dimensionality reduction which attempts to preserve global properties of the data. It maps the high-dimensional data representation to a low-dimensional representation while retaining the pairwise distances between the data points as faithfully as possible.

Suppose for a set of points , . The pairwise Euclidean distance between the data points and is where denotes the th component of . Without loss of generality, we can assume that the ’s are centered.

Notice that the squared pairwise distance is

Let the -dimensional vector be

Then, the squared-distance matrix can be rewritten as where is the -dimensional vector of all ones and .

The mean-centered inner product matrix is defined as where is the mean-centering matrix.

To recover , let the eigendecomposition of be where is the th largest positive eigenvalue and is the corresponding orthonormal eigenvector.

Then, the required -dimensional embedding vectors are given by the columns of the following matrix: where .

MDS is widely used for the visualization of data and in molecular modeling. The popularity of MDS has also led to the proposal of its variants.

##### 2.2. Isomap

When the data points lie on or close to a low-dimensional nonlinear manifold embedded in a high-dimensional space and the nonlinear structure cannot be adequately represented by a linear approximation, classical MDS as mentioned above usually fails to recover the low-dimensional structure of the nonlinear manifold. For example, the data points lying on an underlying manifold two-dimensional “Swiss-roll” embedded in a three-dimensional space cannot be reduced effectively by MDS method.

To deal with this problem, recently, a class of nonlinear embedding techniques has been designed to discover the structure of high-dimensional data and find their embedding in a low-dimensional Euclidean space. Among them, in 2000, Tenenbaum et al. [5] proposed a well-known dimensionality reduction technique, named Isomap, which combines the major algorithmic advantages of PCA and MDS. In the Isomap algorithm, geodesic distances between points are extracted instead of simply taking the Euclidean distance. The geodesic distances are computed by constructing a sparse graph in which each node is connected only to its closest neighbors. The geodesic distance between each pair of nodes is taken to be the length of the shortest path in the graph that connects them. These approximated geodesic distances are then used as input to classical MDS. Up to now, Isomap has proven to be successful in many applications, such as wood inspection, visualization of biomedical data, and head pose estimation.

The Isomap algorithm can be written briefly as follows.

*Step 1 (construct neighborhood graph). *Determine which points are neighbors on the manifold . Two simple methods are to connect each point to all points within some fixed radius (-Isomap) or to all of its nearest neighbors (-Isomap). These neighborhood relations are represented as a weighted graph over the data points, with edges of weight between neighboring points.

*Step 2 (compute shortest paths). *Isomap estimates the geodesic distances between all pairs of points on the manifold by computing their shortest path distance in the graph .

*Step 3 (construct -dimensional embedding). *Apply classical MDS to the matrix of graph distance , constructing an embedding of the data in a low-dimensional Euclidean space that best preserves the estimated intrinsic geometry of the manifold.

-Isomap implies that the method is used in Step 1 in Isomap algorithm. The detailed procedure of Isomap algorithm can be referred to in [5].

#### 3. An Improvement of -Isomap

The basic idea of Isomap algorithm is that, for a neighborhood of points on a manifold, the Euclidean distances provide a fair approximation of geodesic distances. For faraway points the geodesic distance is estimated by the shortest path through neighboring points.

It should be noted that the neighbors satisfy symmetrical relation in -Isomap algorithm. If data point is in the neighbourhood of data point , then must be in the neighbourhood of since the distance between and is less than or equal to . However, the symmetrical relation between neighbors in -Isomap method is no longer right. As shown in Figure 1, if we take nearest neighbors of to construct neighborhood graph in -Isomap, is the fifth neighbor of . But is not one of the five nearest neighbors of .

In this case, it is more reasonable to express the relation between and via although the distance is less than . To best preserve the close relation between neighbors properly in the procedure of dimensionality reduction, a novel technique is described to reduce high-dimensional data into a low-dimensional Euclidean space.

Suppose the data consist of real-valued vectors, each dimensionality , sampled from some underlying manifold. Provided there is sufficient data, we expect each data point and its neighbors to lie on or close to a locally linear patch of the manifold. We assign neighbors to each data point based on the distances between pairs of points and by using the nearest neighbors in the input space. To reflect the nearness of neighbor to data point , is defined as where denotes the distances between neighbor and the data point . If does not belong to the set of neighbors of , then .

It is obvious that is generally not equal to . In order to use MDS algorithm, therefore, we define the symmetrical matrix as follows:

The elements of matrix indicate the average weights between and . The neighborhood relations are represented by a weighted graph over the data points, with edges of weight between neighboring points. The graph constructed by the proposed method is more effective to reflect the neighborhood relations between data points than that defined by -Isomap. For instance, the edge connected data points and in Figure 1 do not exist according to our method because the relation between them is not close together.

The geodesic distances between all pairs of points on the manifold can be obtained by computing their shortest path distances in the graph . As -Isomap algorithm, we apply classical MDS to the matrix , constructing an embedding of the data in a low-dimensional Euclidean space that best preserves the manifolds estimated intrinsic geometry.

#### 4. Experiments

To verify the dimensionality reduction capability of the introduced method, we would like to test it on three classical examples which are widely used to test the effectiveness of dimensionality reduction techniques. The experimental platform is based on Windows 7 with AMD Phenom (tm) II P960 Quad-Core Processor 1.8 GHz and 2.00 GB memory. The programming language is Matlab R2011a.

In what follows, two classical pattern classification problems, face recognition and handwritten digit recognition, are considered in order to analyze the performance of our proposal.

One of the classical examples in dimensionality reduction from the domain of visual perception is face recognition. Multiple photographs () of the same face with different poses and lighting directions are digitized as gray scale images. All of the images which are saved in no particular order can be considered as points in a high-dimensional vector space and lie on an intrinsically three-dimensional manifold. We test the improved method against this dataset (http://isomap.stanford.edu/datasets.html) and take . A two-dimensional projection is shown in Figure 2. The circles indicate the original input images and the points reflect the neighbor relations of the images. The direction of coordinate axis represents the gradual change of poses. The results show that the relations of images with similar poses are preserved perfectly on this dataset in the procedure of dimensionality reduction.

The MNIST database of handwritten digits, available from http://yann.lecun.com/exdb/mnist/, has a training set of 60,000 examples and a test set of 10,000 examples. The digits have been size-normalized and centered in a fixed-size image. Each digit set in the dataset consists of about 1100 instances of a given numeral (“0” to “9”) scanned in as pixel gray scale images. These handwritten digits vary in slope, angle, and thickness characters. In order to evaluate the performance of our proposal, the digits “2” which are written in various forms are selected from the MNIST database. The major features of the “2” are of bottom loop, top arch, and thickness. Applied to this dataset, the proposed method () learns a two-dimensional embedding of the data’s intrinsic geometric structure. It is easy to see from Figure 3 that similar symbols are near each other. This fact indicates that the neighbors’ relation of data points in high-dimensional space still holds in two-dimensional space by using the introduced approach.

A low-dimensional nonlinear manifold embedded in a high-dimensional space is generally invisible, and therefore it is difficult to understand the dimensionality reduction clearly. To illustrate it effectively, we would like to introduce an artificial dataset “Swiss-roll” which is three-dimensional data () sampled from a two-dimensional manifold as shown in Figure 4(a). The colors of data points in Figure 4 indicate their neighborhood relations. It is easy to see from Figure 4(b) that the “Swiss-roll” is unfolded by the proposed approach () in two-dimensional space. The example explains visually that the neighborhood relations of all data points in high-dimensional space are preserved well while reducing them into low-dimensional space by our proposal.

**(a)**

**(b)**

##### 4.1. Discussions and Conclusions

Since two famous methods, Isomap and LLE, were proposed in 2000, the nonlinear techniques for dimensionality reduction have been widely investigated and successfully applied in various scientific fields, such as pattern recognition, visual perception of images of faces, handwriting, and gestures. Along this way, an improvement of -Isomap algorithm is presented to reduce the dataset in high-dimensional space into a new dataset in low-dimensional space. The neighborhood relations can be preserved well in the process of dimensionality reduction by the proposal. The reduced results of three classical examples account for the validity of our proposed method.

It is well known that the dimensionality reduction is a stage of data preprocessing in data mining, not our final destination. It is worthwhile to investigate how to apply the reduced results to classify/cluster the real/artificial datasets. Our future work will focus on this problem.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### References

- L. van der Maaten, E. Postma, and J. van den Herik, “Dimensionality reduction: a comparative review,”
*Journal of Machine Learning Research*, vol. 10, pp. 1–41, 2009. View at: Google Scholar - I. T. Jolliffe,
*Principal Component Analysis*, Springer, New York, NY, USA, 2nd edition, 2002. View at: MathSciNet - S. Mika,
*Kernel fisher discriminant [Ph.D. thesis]*, University of Technology, Berlin, Germany, 2002. - W. S. Torgerson, “Multidimensional scaling: theory and method,”
*Psychometrika*, vol. 17, pp. 401–419, 1952. View at: Google Scholar | Zentralblatt MATH | MathSciNet - J. B. Tenenbaum, V. de Silva, and J. C. Langford, “A global geometric framework for nonlinear dimensionality reduction,”
*Science*, vol. 290, no. 5500, pp. 2319–2323, 2000. View at: Google Scholar - S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by locally linear embedding,”
*Science*, vol. 290, no. 5500, pp. 2323–2326, 2000. View at: Google Scholar - H. Y. Zha and Z. Y. Zhang, “Isometric embedding and continuum ISOMAP,” in
*Proceedings of the 20th International Conference on Machine Learning (ICML '03)*, pp. 864–871, Washington, DC, USA, 2003. View at: Google Scholar - X. Q. Hu, Z. X. Yang, and L. Jing, “An incremental dimensionality reduction method on discriminant information for pattern classification,”
*Pattern Recognition Letters*, vol. 30, pp. 1416–1423, 2009. View at: Google Scholar - Y. L. Zheng, B. Fang, Y. Y. Tang, T. P. Zhang, and R. Z. Liu, “Learning orthogonal projections for Isomap,”
*Neurocomputing*, vol. 103, pp. 149–154, 2013. View at: Google Scholar - Y. D. Bu, F. Q. Chen, and J. C. Pan, “Stellar spectral subclasses classification based on Isomap and SVM,”
*New Astronomy*, vol. 28, pp. 35–43, 2014. View at: Google Scholar - C. Dimeglio, S. Gallón, J.-M. Loubes, and E. Maza, “A robust algorithm for template curve estimation based on manifold embedding,”
*Computational Statistics & Data Analysis*, vol. 70, pp. 373–386, 2014. View at: Publisher Site | Google Scholar | MathSciNet - H. Choi and S. Choi, “Robust kernel Isomap,”
*Pattern Recognition*, vol. 40, pp. 853–862, 2007. View at: Google Scholar - O. Samko, A. D. Marshall, and P. L. Rosin, “Selection of the optimal parameter value for the Isomap algorithm,”
*Pattern Recognition Letters*, vol. 27, pp. 968–979, 2006. View at: Google Scholar - B. Raducanu and F. Dornaika, “A supervised non-linear dimensionality reduction approach for manifold learning,”
*Pattern Recognition*, vol. 45, pp. 2432–2444, 2012. View at: Google Scholar - C. Williams, “On a connection between kernel PCA and metric multidimensional scaling,”
*Machine Learning*, vol. 46, pp. 11–19, 2002. View at: Google Scholar - T. Cox and M. Cox,
*Multidimensional Scaling*, Chapman & Hall, London, UK, 1994.

#### Copyright

Copyright © 2014 Fuding Xie et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.