Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

December 12, 2016 ยท Entered Twilight ยท ๐Ÿ› International Conference on Learning Representations

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: README.md, cifar.py, imagenet.py, requirements.txt, utils.py, visualize-attention.ipynb

Authors Sergey Zagoruyko, Nikos Komodakis arXiv ID 1612.03928 Category cs.CV: Computer Vision Citations 2.9K Venue International Conference on Learning Representations Repository https://github.com/szagoruyko/attention-transfer โญ 1464 Last Checked 1 month ago
Abstract
Attention plays a critical role in human visual experience. Furthermore, it has recently been demonstrated that attention can also play an important role in the context of applying artificial neural networks to a variety of tasks from fields such as computer vision and NLP. In this work we show that, by properly defining attention for convolutional neural networks, we can actually use this type of information in order to significantly improve the performance of a student CNN network by forcing it to mimic the attention maps of a powerful teacher network. To that end, we propose several novel methods of transferring attention, showing consistent improvement across a variety of datasets and convolutional neural network architectures. Code and models for our experiments are available at https://github.com/szagoruyko/attention-transfer
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision