Training Restricted Boltzmann Machines via the Thouless-Anderson-Palmer Free Energy

June 09, 2015 Β· Entered Twilight Β· πŸ› Neural Information Processing Systems

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: CSBP_Solver.m, DESCRIPTION.txt, Demos, HISTORY.txt, README.txt, Subroutines, data

Authors Marylou Gabrié, Eric W. Tramel, Florent Krzakala arXiv ID 1506.02914 Category cond-mat.dis-nn Cross-listed cs.LG, cs.NE, stat.ML Citations 37 Venue Neural Information Processing Systems Repository https://github.com/jeanbarbier/BPCS_common ⭐ 6 Last Checked 6 days ago
Abstract
Restricted Boltzmann machines are undirected neural networks which have been shown to be effective in many applications, including serving as initializations for training deep multi-layer neural networks. One of the main reasons for their success is the existence of efficient and practical stochastic algorithms, such as contrastive divergence, for unsupervised training. We propose an alternative deterministic iterative procedure based on an improved mean field method from statistical physics known as the Thouless-Anderson-Palmer approach. We demonstrate that our algorithm provides performance equal to, and sometimes superior to, persistent contrastive divergence, while also providing a clear and easy to evaluate objective function. We believe that this strategy can be easily generalized to other models as well as to more accurate higher-order approximations, paving the way for systematic improvements in training Boltzmann machines with hidden units.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” cond-mat.dis-nn