Distilling Knowledge from Self-Supervised Teacher by Embedding Graph Alignment

November 23, 2022 ยท Entered Twilight ยท ๐Ÿ› British Machine Vision Conference

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: CLIP, LICENSE, README.md, crd, dataset, distiller_zoo, environment.txt, figure, helper, models, scripts, train_student.py, train_teacher.py

Authors Yuchen Ma, Yanbei Chen, Zeynep Akata arXiv ID 2211.13264 Category cs.CV: Computer Vision Cross-listed cs.LG Citations 9 Venue British Machine Vision Conference Repository https://github.com/yccm/EGA โญ 10 Last Checked 1 month ago
Abstract
Recent advances have indicated the strengths of self-supervised pre-training for improving representation learning on downstream tasks. Existing works often utilize self-supervised pre-trained models by fine-tuning on downstream tasks. However, fine-tuning does not generalize to the case when one needs to build a customized model architecture different from the self-supervised model. In this work, we formulate a new knowledge distillation framework to transfer the knowledge from self-supervised pre-trained models to any other student network by a novel approach named Embedding Graph Alignment. Specifically, inspired by the spirit of instance discrimination in self-supervised learning, we model the instance-instance relations by a graph formulation in the feature embedding space and distill the self-supervised teacher knowledge to a student network by aligning the teacher graph and the student graph. Our distillation scheme can be flexibly applied to transfer the self-supervised knowledge to enhance representation learning on various student networks. We demonstrate that our model outperforms multiple representative knowledge distillation methods on three benchmark datasets, including CIFAR100, STL10, and TinyImageNet. Code is here: https://github.com/yccm/EGA.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision