EM-Fusion: Dynamic Object-Level SLAM with Probabilistic Data Association

April 26, 2019 Β· Entered Twilight Β· πŸ› IEEE International Conference on Computer Vision

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 5.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitattributes, CMakeLists.txt, EXAMPLE.md, LICENSE, README.md, apps, config, eval_co-fusion.sh, eval_tum.sh, images, include, run_exps.sh, src

Authors Michael Strecke, Jârg Stückler arXiv ID 1904.11781 Category cs.CV: Computer Vision Cross-listed cs.RO Citations 84 Venue IEEE International Conference on Computer Vision Repository https://github.com/EmbodiedVision/emfusion ⭐ 56 Last Checked 1 month ago
Abstract
The majority of approaches for acquiring dense 3D environment maps with RGB-D cameras assumes static environments or rejects moving objects as outliers. The representation and tracking of moving objects, however, has significant potential for applications in robotics or augmented reality. In this paper, we propose a novel approach to dynamic SLAM with dense object-level representations. We represent rigid objects in local volumetric signed distance function (SDF) maps, and formulate multi-object tracking as direct alignment of RGB-D images with the SDF representations. Our main novelty is a probabilistic formulation which naturally leads to strategies for data association and occlusion handling. We analyze our approach in experiments and demonstrate that our approach compares favorably with the state-of-the-art methods in terms of robustness and accuracy.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision