Unsupervised learning of phase transitions: from principal component analysis to variational autoencoders
March 07, 2017 Β· Declared Dead Β· π Physical Review E
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Sebastian Johann Wetzel
arXiv ID
1703.02435
Category
cond-mat.stat-mech
Cross-listed
cs.LG,
stat.ML
Citations
348
Venue
Physical Review E
Last Checked
1 month ago
Abstract
We employ unsupervised machine learning techniques to learn latent parameters which best describe states of the two-dimensional Ising model and the three-dimensional XY model. These methods range from principal component analysis to artificial neural network based variational autoencoders. The states are sampled using a Monte-Carlo simulation above and below the critical temperature. We find that the predicted latent parameters correspond to the known order parameters. The latent representation of the states of the models in question are clustered, which makes it possible to identify phases without prior knowledge of their existence or the underlying Hamiltonian. Furthermore, we find that the reconstruction loss function can be used as a universal identifier for phase transitions.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β cond-mat.stat-mech
R.I.P.
π»
Ghosted
π
π
Old Age
Unsupervised Generative Modeling Using Matrix Product States
R.I.P.
π»
Ghosted
Solving Statistical Mechanics Using Variational Autoregressive Networks
R.I.P.
π»
Ghosted
Learning Thermodynamics with Boltzmann Machines
R.I.P.
π»
Ghosted
Information Flows? A Critique of Transfer Entropies
R.I.P.
π»
Ghosted
Phase Transitions in Semidefinite Relaxations
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted