at DataviewInlineApi.eval (plugin:dataview:19027:21)
at evalInContext (plugin:dataview:19028:7)
at asyncEvalInContext (plugin:dataview:19038:32)
at DataviewJSRenderer.render (plugin:dataview:19064:19)
at DataviewJSRenderer.onload (plugin:dataview:18606:14)
at DataviewJSRenderer.load (app://obsidian.md/app.js:1:1182416)
at DataviewApi.executeJs (plugin:dataview:19607:18)
at DataviewCompiler.eval (plugin:digitalgarden:10763:23)
at Generator.next (<anonymous>)
at eval (plugin:digitalgarden:90:61)
Class Notes
Notes pages 15-20 under Singular Values: Geometric Method
Summary
Concentration of singular values for short/fat matrices
Matrix concentration inequality
Geometric method for random matrix analysis
Epsilon nets
Discretizing matrix norms
Existence of good nets
2. Rectangular Matrices
2.3 Spectral Analysis of Rectangular Gaussian Matrices
2.3.1 Singular Values: Geometric Method
Last time: Spectrum of random matrix
and especially
(not actually tru)
from JL thm for a few quadratic forms
This is a [[semi-norm]] - whether or not it is a [[norm]] depends on the set .
+++ WHEN IS IT A NORM?
To be a norm, outer products need to span the space of symmetric matrices
Proposition
Let be a symmetric [[random matrix]] and let be finite. Then
There is a tradeoff between the factor and the error in approximating the .
Thus the problem of accurate approximation becomes one of finding a small set that behaves like a "grid" approximating (or )
We call this approach the geometric method because it reduces the problem of understanding norms of random matrices to a geometric problem instead
2.3.2 Nets, Coverings, and Packings
Discretizing the operator norm
We want to know if the [[restricted inner product]] approximates the [[operator norm]]
-net
Let . is an -covering or an -net of if
see [[epsilon net]]
Lemma
If is an [[epsilon net]] of or , then
(if for the second inequality)
Proof
Let be such that and let such that . Then
Note
++++ interpretation of special
see [[epsilon net restricted inner product bounds the operator norm]]
We want to construct our epsilon nets so we can control their size. It is hard to manually describe nets in terms of their vectors, so we need some framework to define them otherwise.
Non-constructive existence of nets
-packing
Let . is an -packing of if
ie if any two of these balls are sufficiently far apart. ie,
We say that is maximal if is not a packing for all
Proof
Let . Maximality means for some . Let be in this intersection
Then by the [[triangle inequality]] and by construction of our balls above, we get
see [[maximal epsilon packing is also a net]]
We want to show that small nets exist. We have a volumetric bound on the size of packings (including maximal ones), and as we saw above, this means there is a corresponding net. So we have a plan:
This means that there exists a net that is not too large
Proposition
Suppose is an [[epsilon packing]] of . Then
Note
The true value is is
Proof
Since is a packing of , we know
Note in the first line the LHS union is disjoint, which is how we get the second line.
see [[bound for the size of an epsilon packing]]
Corollary
For all , there exists an [[epsilon net]] of with
Proof
Define any maximal packing. The result follows from [[maximal epsilon packing is also a net]] and the size bound follows immediately from the [[bound for the size of an epsilon packing]]
see [[we can find an epsilon net with the bound for an epsilon packing]]
2.3.3 Coarse Non-Asymptotic Bound on Singular Values
Putting everything together, we can get a bound on .
Using the existence not-too-big [[epsilon net]], we can approximate the norm by discretizing and test vectors in the net.
We can then use the first moment method (Taylor) to help get our bound
Theorem
For, let . Define . Then
Note: You can't use 'macro parameter character #' in math mode\lvert \lvert G \rvert \rvert { #2} =\lvert \lvert GG^T \rvert \rvert \leq m+{\cal O}(\sqrt{ dm })
Proof
Let . Then we have:
There exists an [[epsilon net]] of where we our [[bound for the size of an epsilon packing]] yields
We also have the bound (from [[epsilon net restricted inner product bounds the operator norm]])
So
Since , we know . So let . Then
Now, let
You can't use 'macro parameter character #' in math mode\begin{align} \implies &= \lvert {\cal X} \rvert \cdot \mathbb{P}_{g \sim N(0, I_{m})}\left[ \lvert\,\lvert \lvert g \rvert \rvert { #2} -m \,\rvert \geq \frac{t}{2} \right] \\ &\leq 9^d\cdot 2\exp\left( -\frac{1}{8}\min\left\{ \frac{t^2}{4m}, \frac{t}{2} \right\} \right) \\ &\leq 2\exp\left( 3d - \frac{1}{8} \min\left\{ \frac{C^2}{4}d, \underbrace{\frac{C}{2}\sqrt{ dm }}_{d\leq m \implies \sqrt{ dm } \geq d} \right\} \right) \end{align}
And setting and noting that , we get
For , we want to pick big enough that everything is negative.
The proof is similar to the one proving Weyl, except we can use the variational description of the eigenvalues instead of Courant-Fisher. For the first bound, we have
(by just distributing the and by definition of the [[operator norm]]). And, similarly, we have
Corollary
Let be a Gaussian [[random matrix]] with . Then
Important
ie all singular values are {}
Note
This is a good event; we want this to happen. And luckily, we have high probability for this event.
Proof
Recall
First-order Taylor approximation (and concavity of ) yields
If the event for our [[high probability bound for operator norm of difference for Gaussian covariance matrix]] happens, then we have