Very Deep VAEs Generalize Autoregressive Models and Can Outperform Them on Images

November 20, 2020 ยท Entered Twilight ยท ๐Ÿ› International Conference on Learning Representations

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE.md, README.md, data.py, files_to_npy.py, header-image.png, hps.py, setup_cifar10.sh, setup_ffhq1024.sh, setup_ffhq256.sh, setup_imagenet.sh, train.py, train_helpers.py, utils.py, vae.py, vae_helpers.py

Authors Rewon Child arXiv ID 2011.10650 Category cs.LG: Machine Learning Cross-listed cs.CV Citations 386 Venue International Conference on Learning Representations Repository https://github.com/openai/vdvae โญ 451 Last Checked 1 month ago
Abstract
We present a hierarchical VAE that, for the first time, generates samples quickly while outperforming the PixelCNN in log-likelihood on all natural image benchmarks. We begin by observing that, in theory, VAEs can actually represent autoregressive models, as well as faster, better models if they exist, when made sufficiently deep. Despite this, autoregressive models have historically outperformed VAEs in log-likelihood. We test if insufficient depth explains why by scaling a VAE to greater stochastic depth than previously explored and evaluating it CIFAR-10, ImageNet, and FFHQ. In comparison to the PixelCNN, these very deep VAEs achieve higher likelihoods, use fewer parameters, generate samples thousands of times faster, and are more easily applied to high-resolution images. Qualitative studies suggest this is because the VAE learns efficient hierarchical visual representations. We release our source code and models at https://github.com/openai/vdvae.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning