Nonlinear Dimensionality Compression Algorithm

Jump to navigation Jump to search

A Nonlinear Dimensionality Compression Algorithm is a dimensionality reduction algorithm that uses a non-linear model (to solve a dimensionality reduction task).



    • High-dimensional data, meaning data that requires more than two or three dimensions to represent, can be difficult to interpret. One approach to simplification is to assume that the data of interest lies on an embedded non-linear manifold within the higher-dimensional space. If the manifold is of low enough dimension then the data can be visualised in the low dimensional space.
      Below is a summary of some of the important algorithms from the history of manifold learning and nonlinear dimensionality reduction.[1] Many of these non-linear dimensionality reduction methods are related to the linear methods listed below. Non-linear methods can be broadly classified into two groups: those that provide a mapping (either from the high dimensional space to the low dimensional embedding or vice versa), and those that just give a visualisation. In the context of machine learning, mapping methods may be viewed as a preliminary feature extraction step, after which pattern recognition algorithms are applied. Typically those that just give a visualisation are based on proximity data - that is, distance measurements.
  1. John A. Lee, Michel Verleysen, Nonlinear Dimensionality Reduction, Springer, 2007.