Robust Point Cloud Segmentation with Noisy Annotations

December 06, 2022 ยท Entered Twilight ยท ๐Ÿ› IEEE Transactions on Pattern Analysis and Machine Intelligence

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: NL_S3DIS, PNAL_dgcnn.py, README.md, configs, dgcnn_segmentation.py, find_rgb.py, imgs, main, method, pointnet2_classification.py, run.sh, run_pnal.sh, structure, utils

Authors Shuquan Ye, Dongdong Chen, Songfang Han, Jing Liao arXiv ID 2212.03242 Category cs.CV: Computer Vision Cross-listed cs.LG Citations 15 Venue IEEE Transactions on Pattern Analysis and Machine Intelligence Repository https://github.com/pleaseconnectwifi/PNAL โญ 67 Last Checked 1 month ago
Abstract
Point cloud segmentation is a fundamental task in 3D. Despite recent progress on point cloud segmentation with the power of deep networks, current learning methods based on the clean label assumptions may fail with noisy labels. Yet, class labels are often mislabeled at both instance-level and boundary-level in real-world datasets. In this work, we take the lead in solving the instance-level label noise by proposing a Point Noise-Adaptive Learning (PNAL) framework. Compared to noise-robust methods on image tasks, our framework is noise-rate blind, to cope with the spatially variant noise rate specific to point clouds. Specifically, we propose a point-wise confidence selection to obtain reliable labels from the historical predictions of each point. A cluster-wise label correction is proposed with a voting strategy to generate the best possible label by considering the neighbor correlations. To handle boundary-level label noise, we also propose a variant ``PNAL-boundary " with a progressive boundary label cleaning strategy. Extensive experiments demonstrate its effectiveness on both synthetic and real-world noisy datasets. Even with $60\%$ symmetric noise and high-level boundary noise, our framework significantly outperforms its baselines, and is comparable to the upper bound trained on completely clean data. Moreover, we cleaned the popular real-world dataset ScanNetV2 for rigorous experiment. Our code and data is available at https://github.com/pleaseconnectwifi/PNAL.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision