Next: 2.1.1 Neural networks for
Up: 2. Modeling of data
Previous: 2. Modeling of data
2.1 Principal component analysis (PCA)
Principal component analysis is a widely used tool for dimension reduction (Diamantaras and Kung, 1996). Let
IRd, where
i = 1,..., n, be the training patterns. The principal components are a set of q < d orthonormal vectors and span a subspace in the major directions into which the patterns extend (figure 2.1).
Figure 2.1:
The principal component
points into the direction of maximum variance. The gray dots are the training patterns. The intersection of the dashed lines is the center of the pattern distribution.
|
In this section, we assume that the patterns are centered around the origin (without loss of generality). Let
be the projection onto a subspace,
is a d×q matrix that contains the principal components as columns.
The vector
is a dimension-reduced representation of
. Let
be the reconstruction of
given only the vector
,
The goal of PCA is to set the subspace such that the mean reconstruction error
Erec is minimized,
This goal is equivalent to finding the q major directions of maximal variance within the set of patterns
{
} (Diamantaras and Kung, 1996). Moreover, it is equivalent to the principal components being the first q eigenvectors
of the covariance matrix
of the pattern set (Diamantaras and Kung, 1996),
The corresponding eigenvalue equation is
The eigenvalue
is the variance of the distribution
{
} in the direction of
. The following sections describe how neural networks can extract principal components and how PCA can be linked to the probability density of a pattern distribution.
Subsections
Next: 2.1.1 Neural networks for
Up: 2. Modeling of data
Previous: 2. Modeling of data
Heiko Hoffmann
2005-03-22