OPERA: Omni-Supervised Representation Learning with Hierarchical Supervisions

October 11, 2022 ยท Entered Twilight ยท ๐Ÿ› IEEE International Conference on Computer Vision

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .DS_Store, README.md, convert_to_deit.py, figures, main_lincls.py, main_moco.py, moco, timm, transfer, vits.py

Authors Chengkun Wang, Wenzhao Zheng, Zheng Zhu, Jie Zhou, Jiwen Lu arXiv ID 2210.05557 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.LG Citations 5 Venue IEEE International Conference on Computer Vision Repository https://github.com/wangck20/OPERA โญ 34 Last Checked 1 month ago
Abstract
The pretrain-finetune paradigm in modern computer vision facilitates the success of self-supervised learning, which tends to achieve better transferability than supervised learning. However, with the availability of massive labeled data, a natural question emerges: how to train a better model with both self and full supervision signals? In this paper, we propose Omni-suPErvised Representation leArning with hierarchical supervisions (OPERA) as a solution. We provide a unified perspective of supervisions from labeled and unlabeled data and propose a unified framework of fully supervised and self-supervised learning. We extract a set of hierarchical proxy representations for each image and impose self and full supervisions on the corresponding proxy representations. Extensive experiments on both convolutional neural networks and vision transformers demonstrate the superiority of OPERA in image classification, segmentation, and object detection. Code is available at: https://github.com/wangck20/OPERA.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision