almost exact recovery is impossible when the signal to noise ratio is less than the threshold

Created Last Edited Edits
2025-02-10 2025-05-13 3
Data
Theorem

Suppose we have a stochastic block model problem where the adjacency matrix AP,P=YBYT, and B={p if c1=c2q otherwise.

If the signal to noise ratio SNR<1n, then almost exact recovery is impossible.

Even with infinite time and resources, there is no algorithm that can recover the true communities with just the probabilities of adjacency. This gives us the information theoretic threshold

Proof

see proofs in Massoulié (2014) and Mossel (2014). Aside: interesting to see since there are different proof methods from different domains. See also Abbe (survey papers).

Example: Sparse Graphs

p=an,q=bn and |E|n20 as n (the average degree vanishes).
Then $$SNR = \frac{\left( \frac{a}{n} - \frac{b}{n} \right)^2}{2\frac{(a+b)}{n}} = \frac{1}{2n} \frac{(a-b)^2}{(a+b)}$$
ie, in sparse graphs, it is not difficult to identify the information theoretic threshhold (we can easily calculate the signal to noise ratio)

Mentions

File
information theoretic threshold
2025-02-10 graphs lecture 6
2025-02-12 graphs lecture 7