1. Construct the sample variance-covariance matrix
2. Find the eigenvectors
3. Projection : use each eigenvector to form a linear combination of
original variables
4. The larger, the better :
the k-th principal component is the projection with the k-th largest eigenvalue
Alternative view
1-D data matrix : rank 1
2-D data matrix :rank 2
K-D data matrix : rank k
Eigenvectors for 1-D sample covariance matrix: rank 1
Eigenvectors for 2-D sample covariance matrix: rank 2
Eigenvectors for k-D sample matrix
Conclusion :
Principal components can be found by applying eigenvalue
decomposition to the
sample covariance matrix of X .
Further discussion :
Adding i.i.d. noise. This does not change the eigenvectors.
Why ?
(because the new covariance matrix is
equal to the old one plus a multiple of the identity matrix
I )
Connection with automatic basis curve finding (to be discussed later)