Control Variates for Stochastic Gradient MCMC
June 16, 2017 ยท Declared Dead ยท ๐ Statistics and computing
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Jack Baker, Paul Fearnhead, Emily B. Fox, Christopher Nemeth
arXiv ID
1706.05439
Category
stat.CO
Cross-listed
cs.LG,
stat.ML
Citations
106
Venue
Statistics and computing
Last Checked
1 month ago
Abstract
It is well known that Markov chain Monte Carlo (MCMC) methods scale poorly with dataset size. A popular class of methods for solving this issue is stochastic gradient MCMC. These methods use a noisy estimate of the gradient of the log posterior, which reduces the per iteration computational cost of the algorithm. Despite this, there are a number of results suggesting that stochastic gradient Langevin dynamics (SGLD), probably the most popular of these methods, still has computational cost proportional to the dataset size. We suggest an alternative log posterior gradient estimate for stochastic gradient MCMC, which uses control variates to reduce the variance. We analyse SGLD using this gradient estimate, and show that, under log-concavity assumptions on the target distribution, the computational cost required for a given level of accuracy is independent of the dataset size. Next we show that a different control variate technique, known as zero variance control variates can be applied to SGMCMC algorithms for free. This post-processing step improves the inference of the algorithm by reducing the variance of the MCMC output. Zero variance control variates rely on the gradient of the log posterior; we explore how the variance reduction is affected by replacing this with the noisy gradient estimate calculated by SGMCMC.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ stat.CO
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Edward: A library for probabilistic modeling, inference, and criticism
R.I.P.
๐ป
Ghosted
Coresets for Scalable Bayesian Logistic Regression
R.I.P.
๐ป
Ghosted
colorspace: A Toolbox for Manipulating and Assessing Colors and Palettes
R.I.P.
๐ป
Ghosted
Fast Discrete Distribution Clustering Using Wasserstein Barycenter with Sparse Support
R.I.P.
๐ป
Ghosted
Poisson multi-Bernoulli conjugate prior for multiple extended object filtering
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted