<< Chapter < Page Chapter >> Page >
Manifold learning demonstration. (a) As input to the manifold learning algorithm, 1000 images of size 64 × 64 are created, where each image consists of a white disk translated to a random position ( θ 1 , θ 2 ) . It follows that the images represent a sampling of 1000 points from a 2 -dimensional submanifold of R 4096 . (b) Scatter plot of the true values for the ( θ 1 , θ 2 ) positions. For visibility in each plot, the color of each point indicates the true θ 1 value. (c) ISOMAP embedding learned from original data points in R 4096 . From the low-dimensional embedding coordinates we can infer the relative positions of the original high-dimensional images. (d) ISOMAP embedding learned from a random projection of the data set to R M , where M = 15 .

These algorithms can be useful for learning the dimension and parametrizations of manifolds, forsorting data, for visualization and navigation through the data, and as preprocessing to make further analysis more tractable;common demonstrations include analysis of face images and classification of and handwritten digits. A related technique, theWhitney Reduction Network [link] , [link] , seeks a linear mapping to R M that preserves ambient pairwise distances on the manifold and is particularly useful forprocessing the output of dynamical systems having low-dimensional attractors.

Other algorithms have been proposed for characterizing manifolds from sampled data without constructing an explicit embedding in R M . The Geodesic Minimal Spanning Tree (GMST)  [link] models the data as random samples from the manifold and estimates the corresponding entropy anddimensionality. Another technique  [link] has been proposed for using random samples of a manifold to estimate its homology(via the Betti numbers, which essentially characterize its dimension, number of connected components, etc.). PersistenceBarcodes  [link] are a related technique that involves constructing a type of signature for a manifold (orsimply a shape) that uses tangent complexes to detect and characterize local edges and corners.

Additional algorithms have been proposed for constructing meaningful functions on the point samples in R N . To solve a semi-supervised learning problem, a method calledLaplacian Eigenmaps  [link] has been proposed that involves forming an adjacency graph for the data in R N , computing eigenfunctions of the Laplacian operator on the graph (which forma basis for L 2 on the graph), and using these functions to train a classifier on the data. The resulting classifiers havebeen used for handwritten digit recognition, document classification, and phoneme classification. (The M smoothest eigenfunctions can also be used to embed the manifold in M , similar to the approaches described above.) A related methodcalled Diffusion Wavelets  [link] uses powers of the diffusion operator to model scale on the manifold, then constructswavelets to capture local behavior at each scale. The result is a wavelet transform adapted not to geodesic distance but todiffusion distance, which measures (roughly) the number of paths connecting two points.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Concise signal models. OpenStax CNX. Sep 14, 2009 Download for free at http://cnx.org/content/col10635/1.4
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Concise signal models' conversation and receive update notifications?

Ask
Briana Knowlton
Start Quiz
Jordon Humphreys
Start Quiz
Mike Wolf
Start Exam
Royalle Moore
Start Quiz
Sarah Warren
Start Test