DeepLiDARFlow: A Deep Learning Architecture For Scene Flow Estimation Using Monocular Camera and Sparse LiDAR

August 18, 2020 Β· Entered Twilight Β· πŸ› IEEE/RJS International Conference on Intelligent RObots and Systems

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 5.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE.md, README.md, data_loader, deeplidarflow.py, download_models.sh, download_preprocessed_kitti.sh, evaluation.py, images, inference.py, modules.py, networks.py, requirements.txt, run_evaluation.sh, run_finetune.sh, run_inference.sh, run_train.sh, train.py, utils

Authors Rishav, Ramy Battrawy, René Schuster, Oliver Wasenmüller, Didier Stricker arXiv ID 2008.08136 Category cs.CV: Computer Vision Cross-listed cs.LG, cs.RO Citations 35 Venue IEEE/RJS International Conference on Intelligent RObots and Systems Repository https://github.com/dfki-av/DeepLiDARFlow ⭐ 5 Last Checked 1 month ago
Abstract
Scene flow is the dense 3D reconstruction of motion and geometry of a scene. Most state-of-the-art methods use a pair of stereo images as input for full scene reconstruction. These methods depend a lot on the quality of the RGB images and perform poorly in regions with reflective objects, shadows, ill-conditioned light environment and so on. LiDAR measurements are much less sensitive to the aforementioned conditions but LiDAR features are in general unsuitable for matching tasks due to their sparse nature. Hence, using both LiDAR and RGB can potentially overcome the individual disadvantages of each sensor by mutual improvement and yield robust features which can improve the matching process. In this paper, we present DeepLiDARFlow, a novel deep learning architecture which fuses high level RGB and LiDAR features at multiple scales in a monocular setup to predict dense scene flow. Its performance is much better in the critical regions where image-only and LiDAR-only methods are inaccurate. We verify our DeepLiDARFlow using the established data sets KITTI and FlyingThings3D and we show strong robustness compared to several state-of-the-art methods which used other input modalities. The code of our paper is available at https://github.com/dfki-av/DeepLiDARFlow.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision