Accelerating Diffusion Sampling with Classifier-based Feature Distillation

November 22, 2022 ยท Declared Dead ยท ๐Ÿ› IEEE International Conference on Multimedia and Expo

๐Ÿ’€ CAUSE OF DEATH: 404 Not Found
Code link is broken/dead
Authors Wujie Sun, Defang Chen, Can Wang, Deshi Ye, Yan Feng, Chun Chen arXiv ID 2211.12039 Category cs.CV: Computer Vision Citations 19 Venue IEEE International Conference on Multimedia and Expo Repository https://github.com/zju-SWJ/RCFD} Last Checked 1 month ago
Abstract
Although diffusion model has shown great potential for generating higher quality images than GANs, slow sampling speed hinders its wide application in practice. Progressive distillation is thus proposed for fast sampling by progressively aligning output images of $N$-step teacher sampler with $N/2$-step student sampler. In this paper, we argue that this distillation-based accelerating method can be further improved, especially for few-step samplers, with our proposed \textbf{C}lassifier-based \textbf{F}eature \textbf{D}istillation (CFD). Instead of aligning output images, we distill teacher's sharpened feature distribution into the student with a dataset-independent classifier, making the student focus on those important features to improve performance. We also introduce a dataset-oriented loss to further optimize the model. Experiments on CIFAR-10 show the superiority of our method in achieving high quality and fast sampling. Code is provided at \url{https://github.com/zju-SWJ/RCFD}.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision

Died the same way โ€” ๐Ÿ’€ 404 Not Found