Principal component analysis: Difference between revisions

Content deleted Content added
m temporarily removing theorem to rewrite
m and its back
Line 156:
 
 
<blockquote>
'''Theorem (Optimal k‑dimensional fit).'''
Let P be an n×m data matrix whose columns have been mean‑centered and scaled, and let
<math>P = U \,\Sigma\, V^{T}</math>
be its singular value decomposition. Then the best rank‑k approximation to P in the least‑squares (Frobenius‑norm) sense is
<math>P_{k} = U_{k}\,\Sigma_{k}\,V_{k}^{T}</math>,
where V<sub>k</sub> consists of the first k columns of V. Moreover, the relative residual variance is
<math>R(k)=\frac{\sum_{j=k+1}^{m}\sigma_{j}^{2}}{\sum_{j=1}^{m}\sigma_{j}^{2}}</math>.
</blockquote>
 
<ref>{{cite book
| last = Holmes
| first = M.
| title = Introduction to Scientific Computing and Data Analysis
| edition = 2nd
| year = 2023
| publisher = Springer
| isbn = 978-3-031-22429-4
}}</ref>