Visual Dynamics: Stochastic Future Generation via Layered Cross Convolutional Networks

July 24, 2018 ยท Entered Twilight ยท ๐Ÿ› IEEE Transactions on Pattern Analysis and Machine Intelligence

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, data, download_models.sh, src

Authors Tianfan Xue, Jiajun Wu, Katherine L. Bouman, William T. Freeman arXiv ID 1807.09245 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.LG Citations 37 Venue IEEE Transactions on Pattern Analysis and Machine Intelligence Repository https://github.com/tfxue/visual-dynamics โญ 74 Last Checked 6 days ago
Abstract
We study the problem of synthesizing a number of likely future frames from a single input image. In contrast to traditional methods that have tackled this problem in a deterministic or non-parametric way, we propose to model future frames in a probabilistic manner. Our probabilistic model makes it possible for us to sample and synthesize many possible future frames from a single input image. To synthesize realistic movement of objects, we propose a novel network structure, namely a Cross Convolutional Network; this network encodes image and motion information as feature maps and convolutional kernels, respectively. In experiments, our model performs well on synthetic data, such as 2D shapes and animated game sprites, and on real-world video frames. We present analyses of the learned network representations, showing it is implicitly learning a compact encoding of object appearance and motion. We also demonstrate a few of its applications, including visual analogy-making and video extrapolation.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision