Lecture 20

[[lecture-data]]

2024-10-14

Readings

7. Chapter 7

Recall from singular value decomposition we have that for any matrix AMm,n with mn

A=UΣW

for some UMm unitary, ΣMm diagonal, and WMm,n with orthogonal columns. Recall that we call the entries of Σ the singular values

Note

The values in Σ for A=UΣW are uniquely determined.

Proof

AA=UΣWWΣU and so the eigenvalues of AA are the diagonal entries of ΣΣ.

Theorem (Singular Value Decomposition)

For any AMm,n, there exists UMm, VMn both unitary and ΣMm,n "diagonal" with min{m,n} real, nonnegative, nonincreasing, entries such that

A=UΣV

Further, if A is real-valued, we can take each of U,Σ,V to be real.

(we take "diagonal" to mean that the only entries that can be nonzero have the same row and column index)

Note

  • We can think of this as a generalization of diagonalization/spectral decomposition, but where the loss is that U and V are distinct.
  • if A is positive semidefinite, then the diagonalization is the singular value decomposition since we have that A=UΣU for some unitary U and nonnegative Σ all real.

Proof

If mn, then by the previous definition here we have A=UΣW where UMn unitary, ΣMm diagonal, and WMm,n with orthogonal columns. Since mn, we add nm columns of zeros to Σ to make it size m×n. Then we can append nm orthogonal rows to W with Gram-Schmidt to get a V

If n>m, then A=UΣV is SVD by the above case. But then A=VΣTU is an SVD of A!

(see singular value decomposition)

Corollary - Polar Decomposition

For all AMn, there exists a hermitian, positive semidefinite PMn and a WMn unitary such that

A=PW
Proof

When n=1, we can write a=αeiθ where α is a PSD 1x1 matrix and eiθ is some unitary matrix (it is a complex number with modulus 1!).

Let A=UΣV be its singular value decomposition. Then

A=UΣUUV=(UΣU)(UV)=PW

(see polar decomposition)

Notes on SVD

Say AMm,n and AUΣV its SVD. Suppose there are k non-negative singular values call them σ1,,σk. Denote u1,um the columns of U and v1,,vn the rows of V. Then

  1. A=UΣV=i=1kσiuivi
  2. rank(A)=rank(Σ)=k
  3. range(A)=span{u1,,uk} - easy to see from (1)
  4. null(A)=span(vk+1,,vn) - easy to see from (1)
  5. range(A)=span{v1,,vk}
  6. null(A)=span{uk+1,,um}
  7. For any AMm,n, we have ||A||2,2:=maxxCn0||Ax||2||x||2=σ1
  8. ||A||F2=||UΣV||F2=||Σ||F2=i=1kσi

To see (5) and (6) , recall that A=VΣTU and we simply apply (3) and (4) in this case.

To see (7), note that

||A||2,22=maxx0||Ax||2||x||2=maxxAAxxx=,σ12
  • Note is by Rayleigh-Ritz that this equals the largest eigenvalue of AA.
  • Then is by the fact that the eigenvalues of AA=(UΣV)(UΣV)=VΣΣV - ie the eigenvalues are the squared singular values of A.

recall that range(A)={x:Ay=x}

Generalized Inverse

Let AMm,n and BMn,m. Let C={1,2,3,4}. Then B is a generalized inverse of A precisely when

For each iC, we call B an "i-generalized inverse"

  • if B satisfies 1, 2, then B is a 1-2-generalized inverse
Note

If A is invertible, then B is a 1-2-3-4-generalized inverse.

Note

0 is a 2-3-4 generalized inverse for every matrix...

Note

If B is a 1-3 for A, then A is a 2-4 for B.

(see generalized inverse)

Proposition

Consider Ax=b. If A is invertible, then x=A1b 🙂

But if A is not square? What do we do? Suppose B is a 1-generalized inverse of A.

Let AMm,n,bCm, and suppose B is a 1-generalized inverse of A. Then if Ax=b is consistent, then x=Bb is a solution.

Proof

Say Az=b ie there is a solution for some zCn. Then

A(Bb)=ABAz=Az=b

(see 1-generalized inverses give solutions to consistent linear systems)