Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
December 12, 2016 ยท Entered Twilight ยท ๐ International Conference on Learning Representations
"Last commit was 7.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: README.md, cifar.py, imagenet.py, requirements.txt, utils.py, visualize-attention.ipynb
Authors
Sergey Zagoruyko, Nikos Komodakis
arXiv ID
1612.03928
Category
cs.CV: Computer Vision
Citations
2.9K
Venue
International Conference on Learning Representations
Repository
https://github.com/szagoruyko/attention-transfer
โญ 1464
Last Checked
1 month ago
Abstract
Attention plays a critical role in human visual experience. Furthermore, it has recently been demonstrated that attention can also play an important role in the context of applying artificial neural networks to a variety of tasks from fields such as computer vision and NLP. In this work we show that, by properly defining attention for convolutional neural networks, we can actually use this type of information in order to significantly improve the performance of a student CNN network by forcing it to mimic the attention maps of a powerful teacher network. To that end, we propose several novel methods of transferring attention, showing consistent improvement across a variety of datasets and convolutional neural network architectures. Code and models for our experiments are available at https://github.com/szagoruyko/attention-transfer
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computer Vision
๐
๐
Old Age
๐
๐
Old Age
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
R.I.P.
๐ป
Ghosted
You Only Look Once: Unified, Real-Time Object Detection
๐
๐
Old Age
SSD: Single Shot MultiBox Detector
๐
๐
Old Age
Squeeze-and-Excitation Networks
R.I.P.
๐ป
Ghosted