DeepVoxels: Learning Persistent 3D Feature Embeddings

December 03, 2018 Β· Entered Twilight Β· πŸ› Computer Vision and Pattern Recognition

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 6.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, .gitmodules, LICENSE, README.md, colmap_wrapper.py, custom_layers.py, data_util.py, dataio.py, deep_voxels.py, environment.yml, losses.py, projection.py, pytorch_prototyping, run_deepvoxels.py, util.py

Authors Vincent Sitzmann, Justus Thies, Felix Heide, Matthias Nießner, Gordon Wetzstein, Michael Zollhâfer arXiv ID 1812.01024 Category cs.CV: Computer Vision Citations 715 Venue Computer Vision and Pattern Recognition Repository https://github.com/vsitzmann/deepvoxels ⭐ 233 Last Checked 1 month ago
Abstract
In this work, we address the lack of 3D understanding of generative neural networks by introducing a persistent 3D feature embedding for view synthesis. To this end, we propose DeepVoxels, a learned representation that encodes the view-dependent appearance of a 3D scene without having to explicitly model its geometry. At its core, our approach is based on a Cartesian 3D grid of persistent embedded features that learn to make use of the underlying 3D scene structure. Our approach combines insights from 3D geometric computer vision with recent advances in learning image-to-image mappings based on adversarial loss functions. DeepVoxels is supervised, without requiring a 3D reconstruction of the scene, using a 2D re-rendering loss and enforces perspective and multi-view geometry in a principled manner. We apply our persistent 3D scene representation to the problem of novel view synthesis demonstrating high-quality results for a variety of challenging scenes.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision