The basic idea of nonlinear component analysis (NCA) or kernel PCA
[60] is to replace the covariance matrix in equation
() with
The mapping
is typically not a surjection, i.e. onto and therefore
the reconstruction of a data vector given the extracted components can be
problematic. Linear principal components can be visualised easily
since they correspond to vectors or directions in the data space,
whereas the kernel-based principal components do not have a
counterpart in the data space. Therefore NCA is not viewed as
a generative model.
Schölkopf et al. [60] have developed an efficient algorithm for NCA where the dimensionality of the feature space can be very large. They also developed a method for iteratively finding a reconstruction in the data space for a point in the feature space. Their experiments suggest that compared to PCA, NCA extracts features which are more useful for classification purposes. The same approach can be used to construct e.g. nonlinear ICA.