Copy-Pasting Coherent Depth Regions Improves Contrastive Learning for Urban-Scene Segmentation

November 25, 2022 Β· Entered Twilight Β· πŸ› British Machine Vision Conference

πŸ’€ TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: MScThesis.pdf, README.md, __init__.py, dataloader, detectron2_model, eval_unsupervised.py, fig, model.py, pixel_grouping, run.py, trainer.py, utils.py

Authors Liang Zeng, Attila Lengyel, Nergis Tâmen, Jan van Gemert arXiv ID 2211.14074 Category cs.CV: Computer Vision Citations 0 Venue British Machine Vision Conference Repository https://github.com/LeungTsang/CPCDR ⭐ 2 Last Checked 1 month ago
Abstract
In this work, we leverage estimated depth to boost self-supervised contrastive learning for segmentation of urban scenes, where unlabeled videos are readily available for training self-supervised depth estimation. We argue that the semantics of a coherent group of pixels in 3D space is self-contained and invariant to the contexts in which they appear. We group coherent, semantically related pixels into coherent depth regions given their estimated depth and use copy-paste to synthetically vary their contexts. In this way, cross-context correspondences are built in contrastive learning and a context-invariant representation is learned. For unsupervised semantic segmentation of urban scenes, our method surpasses the previous state-of-the-art baseline by +7.14% in mIoU on Cityscapes and +6.65% on KITTI. For fine-tuning on Cityscapes and KITTI segmentation, our method is competitive with existing models, yet, we do not need to pre-train on ImageNet or COCO, and we are also more computationally efficient. Our code is available on https://github.com/LeungTsang/CPCDR
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision