Sampling from the Sherrington-Kirkpatrick Gibbs measure via algorithmic stochastic localization
March 10, 2022 Β· Declared Dead Β· π IEEE Annual Symposium on Foundations of Computer Science
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Ahmed El Alaoui, Andrea Montanari, Mark Sellke
arXiv ID
2203.05093
Category
math.PR
Cross-listed
cond-mat.dis-nn,
cs.DS
Citations
92
Venue
IEEE Annual Symposium on Foundations of Computer Science
Last Checked
1 month ago
Abstract
We consider the Sherrington-Kirkpatrick model of spin glasses at high-temperature and no external field, and study the problem of sampling from the Gibbs distribution $ΞΌ$ in polynomial time. We prove that, for any inverse temperature $Ξ²<1/2$, there exists an algorithm with complexity $O(n^2)$ that samples from a distribution $ΞΌ^{alg}$ which is close in normalized Wasserstein distance to $ΞΌ$. Namely, there exists a coupling of $ΞΌ$ and $ΞΌ^{alg}$ such that if $(x,x^{alg})\in\{-1,+1\}^n\times \{-1,+1\}^n$ is a pair drawn from this coupling, then $n^{-1}\mathbb E\{||x-x^{alg}||_2^2\}=o_n(1)$. The best previous results, by Bauerschmidt and Bodineau and by Eldan, Koehler, and Zeitouni, implied efficient algorithms to approximately sample (under a stronger metric) for $Ξ²<1/4$. We complement this result with a negative one, by introducing a suitable "stability" property for sampling algorithms, which is verified by many standard techniques. We prove that no stable algorithm can approximately sample for $Ξ²>1$, even under the normalized Wasserstein metric. Our sampling method is based on an algorithmic implementation of stochastic localization, which progressively tilts the measure $ΞΌ$ towards a single configuration, together with an approximate message passing algorithm that is used to approximate the mean of the tilted measure.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β math.PR
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
An Introduction to Matrix Concentration Inequalities
R.I.P.
π»
Ghosted
Non-backtracking spectrum of random graphs: community detection and non-regular Ramanujan graphs
R.I.P.
π»
Ghosted
Convergence of the Deep BSDE Method for Coupled FBSDEs
R.I.P.
π»
Ghosted
A Random Matrix Approach to Neural Networks
R.I.P.
π»
Ghosted
Concentration and regularization of random graphs
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted