PCA

  • m dim affine hyperplace spanned by first m eigenvectors. Only manifolds and no codebook vectors
  • Be able to reconstruct x from f(x) : decoding function

Steps

  1. Center data (A)
    • Subtract their mean from each pattern.
    • and getting patterns
    • Point Cloud with center of Gravity : origin
      • Extend more in some “directions” characterized by unit norm direction vectors .
      • Distance of a point from the origin in the direction of u : projection of on u aka inner product
      • Extension of cloud in direction u : Mean square dist to origin.
      • Largest extension :
      • Since centered: mean is 0 and is the variance
      • is the longest direction : First PC : PC1
  2. Project points (B)
    • Find orthogonal (90deg) subspace . (n-1) dim linear
    • Map all points to - Second PC : PC2
  3. Rinse and repeat (C)
  4. New PCs plotted in original cloud (D)
  5. For featurres ,
  6. Reconstruction :
    • First few PCs till index m
      • Decoding function
  • How good is the reconstruction
    • Relative amount of dissimilarity to mean empirical variance of patterns - 1
      • Ratio very small as index k grows. Very little info lost by reducing dims. Aka good for very high dim stuff.
  1. Compute SVD
    • form orthonormal, real eigenvectors
    • variances are eigenvalues
    • to get PC vectors lined up in U and variances as eigenvalues in
    • If we want to preserve 98% variance : Rhs of (1) st. ratio is (1-0.98)