Train Sparsely, Generate Densely: Memory-efficient Unsupervised Training of High-resolution Temporal GAN

November 22, 2018 ยท Entered Twilight ยท ๐Ÿ› International Journal of Computer Vision

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .dockerignore, .gitignore, LICENSE, README.md, conf, docker, exp, figure, images, requirements.txt, scripts, setup.cfg, tgan2, train.py

Authors Masaki Saito, Shunta Saito, Masanori Koyama, Sosuke Kobayashi arXiv ID 1811.09245 Category cs.CV: Computer Vision Citations 159 Venue International Journal of Computer Vision Repository https://github.com/pfnet-research/tgan2 โญ 83 Last Checked 1 month ago
Abstract
Training of Generative Adversarial Network (GAN) on a video dataset is a challenge because of the sheer size of the dataset and the complexity of each observation. In general, the computational cost of training GAN scales exponentially with the resolution. In this study, we present a novel memory efficient method of unsupervised learning of high-resolution video dataset whose computational cost scales only linearly with the resolution. We achieve this by designing the generator model as a stack of small sub-generators and training the model in a specific way. We train each sub-generator with its own specific discriminator. At the time of the training, we introduce between each pair of consecutive sub-generators an auxiliary subsampling layer that reduces the frame-rate by a certain ratio. This procedure can allow each sub-generator to learn the distribution of the video at different levels of resolution. We also need only a few GPUs to train a highly complex generator that far outperforms the predecessor in terms of inception scores.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision