singular value decomposition

Data

subject:: Matrix Analysis
parent:: Chapter 7
theme:: math notes

[[dg-publish]] 

Definitions

Theorem (Matrix Analysis 1)

Theorem (Matrix Analysis 2)

Theorem (Singular Value Decomposition)

For any AMm,n, there exists UMm, VMn both unitary and ΣMm,n "diagonal" with min{m,n} real, nonnegative, nonincreasing, entries such that

A=UΣV

Further, if A is real-valued, we can take each of U,Σ,V to be real.

(we take "diagonal" to mean that the only entries that can be nonzero have the same row and column index)

Note

  • We can think of this as a generalization of diagonalization/spectral decomposition, but where the loss is that U and V are distinct.
  • if A is positive semidefinite, then the diagonalization is the singular value decomposition since we have that A=UΣU for some unitary U and nonnegative Σ all real.

Proof

If mn, then by the previous definition here we have A=UΣW where UMn unitary, ΣMm diagonal, and WMm,n with orthogonal columns. Since mn, we add nm columns of zeros to Σ to make it size m×n. Then we can append nm orthogonal rows to W with Gram-Schmidt to get a V

If n>m, then A=UΣV is SVD by the above case. But then A=VΣTU is an SVD of A!

Another Definition (Data Science)

X=UΣV

Where UCn×n,VCm×m are unitary and orthonormal. A unitary matrix U is one such that UU=UU=I. The columns of U are called the left singular vectors of X, and the columns of V are called the right singular vectors. The entries of Σ are called the singular values

Notes on SVD

Say AMm,n and AUΣV its SVD. Suppose there are k non-negative singular values call them σ1,,σk. Denote u1,um the columns of U and v1,,vn the rows of V. Then

  1. A=UΣV=i=1kσiuivi
  2. rank(A)=rank(Σ)=k
  3. range(A)=span{u1,,uk} - easy to see from (1)
  4. null(A)=span(vk+1,,vn) - easy to see from (1)
  5. range(A)=span{v1,,vk}
  6. null(A)=span{uk+1,,um}
  7. For any AMm,n, we have ||A||2,2:=maxxCn0||Ax||2||x||2=σ1
  8. ||A||F2=||UΣV||F2=||Σ||F2=i=1kσi

To see (5) and (6) , recall that A=VΣTU and we simply apply (3) and (4) in this case.

To see (7), note that

||A||2,22=maxx0||Ax||2||x||2=maxxAAxxx=,σ12
  • Note is by Rayleigh-Ritz that this equals the largest eigenvalue of AA.
  • Then is by the fact that the eigenvalues of AA=(UΣV)(UΣV)=VΣΣV - ie the eigenvalues are the squared singular values of A.

Mentions

File
Moore-Penrose inverse
Notes on SVD - from Brunton and Kutz
Procrustes Notes
polar decomposition
singular value decomposition
the psuedoinverse gives the least norm solution to the least squares problem
Lecture 19
Lecture 20
Lecture 21
Lecture 22
Lecture 26
Lecture 36
Section 09
my obsidian vault