Flow++: Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

February 01, 2019 ยท Entered Twilight ยท ๐Ÿ› International Conference on Machine Learning

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 6.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE.md, README.md, flows, flows_celeba, flows_imagenet

Authors Jonathan Ho, Xi Chen, Aravind Srinivas, Yan Duan, Pieter Abbeel arXiv ID 1902.00275 Category cs.LG: Machine Learning Cross-listed cs.NE, stat.ML Citations 488 Venue International Conference on Machine Learning Repository https://github.com/aravindsrinivas/flowpp โญ 191 Last Checked 1 month ago
Abstract
Flow-based generative models are powerful exact likelihood models with efficient sampling and inference. Despite their computational efficiency, flow-based models generally have much worse density modeling performance compared to state-of-the-art autoregressive models. In this paper, we investigate and improve upon three limiting design choices employed by flow-based models in prior work: the use of uniform noise for dequantization, the use of inexpressive affine flows, and the use of purely convolutional conditioning networks in coupling layers. Based on our findings, we propose Flow++, a new flow-based model that is now the state-of-the-art non-autoregressive model for unconditional density estimation on standard image benchmarks. Our work has begun to close the significant performance gap that has so far existed between autoregressive models and flow-based models. Our implementation is available at https://github.com/aravindsrinivas/flowpp
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning