Channel-Recurrent Autoencoding for Image Modeling

June 12, 2017 ยท Entered Twilight ยท ๐Ÿ› arXiv.org

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: KLDCriterion.lua, README.md, RNNinit.lua, Sampler.lua, adam_gan.lua, generate_images.lua, image.lua, image_utils.lua, init.lua, main.lua, main_mnist.lua, models, opts.lua, train_test, utils

Authors Wenling Shang, Kihyuk Sohn, Yuandong Tian arXiv ID 1706.03729 Category cs.LG: Machine Learning Cross-listed cs.CV Citations 3 Venue arXiv.org Repository https://github.com/WendyShang/crVAE โญ 5 Last Checked 1 month ago
Abstract
Despite recent successes in synthesizing faces and bedrooms, existing generative models struggle to capture more complex image types, potentially due to the oversimplification of their latent space constructions. To tackle this issue, building on Variational Autoencoders (VAEs), we integrate recurrent connections across channels to both inference and generation steps, allowing the high-level features to be captured in global-to-local, coarse-to-fine manners. Combined with adversarial loss, our channel-recurrent VAE-GAN (crVAE-GAN) outperforms VAE-GAN in generating a diverse spectrum of high resolution images while maintaining the same level of computational efficacy. Our model produces interpretable and expressive latent representations to benefit downstream tasks such as image completion. Moreover, we propose two novel regularizations, namely the KL objective weighting scheme over time steps and mutual information maximization between transformed latent variables and the outputs, to enhance the training.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning