Information-theoretic thresholds for community detection in sparse networks
January 11, 2016 Β· Declared Dead Β· π Annual Conference Computational Learning Theory
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Jess Banks, Cristopher Moore
arXiv ID
1601.02658
Category
math.PR
Cross-listed
cond-mat.stat-mech,
cs.CC,
cs.IT,
cs.SI
Citations
128
Venue
Annual Conference Computational Learning Theory
Last Checked
1 month ago
Abstract
We give upper and lower bounds on the information-theoretic threshold for community detection in the stochastic block model. Specifically, let $k$ be the number of groups, $d$ be the average degree, the probability of edges between vertices within and between groups be $c_\mathrm{in}/n$ and $c_\mathrm{out}/n$ respectively, and let $Ξ»= (c_\mathrm{in}-c_\mathrm{out})/(kd)$. We show that, when $k$ is large, and $Ξ»= O(1/k)$, the critical value of $d$ at which community detection becomes possible -- in physical terms, the condensation threshold -- is \[ d_c = Ξ\!\left( \frac{\log k}{k Ξ»^2} \right) \, , \] with tighter results in certain regimes. Above this threshold, we show that the only partitions of the nodes into $k$ groups are correlated with the ground truth, giving an exponential-time algorithm that performs better than chance -- in particular, detection is possible for $k \ge 5$ in the disassortative case $Ξ»< 0$ and for $k \ge 11$ in the assortative case $Ξ»> 0$. (Similar upper bounds were obtained independently by Abbe and Sandon.) Below this threshold, we use recent results of Neeman and Netrapalli (who generalized arguments of Mossel, Neeman, and Sly) to show that no algorithm can label the vertices better than chance, or even distinguish the block model from an ErdΕs-RΓ©nyi random graph with high probability. We also rely on bounds on certain functions of doubly stochastic matrices due to Achlioptas and Naor; indeed, our lower bound on $d_c$ is the second moment lower bound on the $k$-colorability threshold for random graphs with a certain effective degree.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β math.PR
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
An Introduction to Matrix Concentration Inequalities
R.I.P.
π»
Ghosted
Non-backtracking spectrum of random graphs: community detection and non-regular Ramanujan graphs
R.I.P.
π»
Ghosted
Convergence of the Deep BSDE Method for Coupled FBSDEs
R.I.P.
π»
Ghosted
A Random Matrix Approach to Neural Networks
R.I.P.
π»
Ghosted
Concentration and regularization of random graphs
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted