3DMatch: Learning Local Geometric Descriptors from RGB-D Reconstructions

March 27, 2016 Β· Entered Twilight Β· πŸ› Computer Vision and Pattern Recognition

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, BuildCorrespondence, ElasticReconstruction.sln, FragmentOptimizer, GlobalRegistration, GraphOptimizer, Integrate, LICENSE.txt, Matlab_Toolbox, README.txt

Authors Andy Zeng, Shuran Song, Matthias Nießner, Matthew Fisher, Jianxiong Xiao, Thomas Funkhouser arXiv ID 1603.08182 Category cs.CV: Computer Vision Citations 1.1K Venue Computer Vision and Pattern Recognition Repository https://github.com/qianyizh/ElasticReconstruction ⭐ 660 Last Checked 6 days ago
Abstract
Matching local geometric features on real-world depth images is a challenging task due to the noisy, low-resolution, and incomplete nature of 3D scan data. These difficulties limit the performance of current state-of-art methods, which are typically based on histograms over geometric properties. In this paper, we present 3DMatch, a data-driven model that learns a local volumetric patch descriptor for establishing correspondences between partial 3D data. To amass training data for our model, we propose a self-supervised feature learning method that leverages the millions of correspondence labels found in existing RGB-D reconstructions. Experiments show that our descriptor is not only able to match local geometry in new scenes for reconstruction, but also generalize to different tasks and spatial scales (e.g. instance-level object model alignment for the Amazon Picking Challenge, and mesh surface correspondence). Results show that 3DMatch consistently outperforms other state-of-the-art approaches by a significant margin. Code, data, benchmarks, and pre-trained models are available online at http://3dmatch.cs.princeton.edu
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision