Class-Weighted Convolutional Features for Visual Instance Search

July 09, 2017 ยท Entered Twilight ยท ๐Ÿ› British Machine Vision Conference

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, _config.yml, authors, figs, jimenez-bmvc-2017-paper.pdf, jimenez-bmvc-2017-poster.pdf, keras_code, lists, logos, pytorch_code, requirements.txt, slides

Authors Albert Jimenez, Jose M. Alvarez, Xavier Giro-i-Nieto arXiv ID 1707.02581 Category cs.CV: Computer Vision Cross-listed cs.IR, cs.MM Citations 81 Venue British Machine Vision Conference Repository https://github.com/imatge-upc/retrieval-2017-cam โญ 224 Last Checked 8 days ago
Abstract
Image retrieval in realistic scenarios targets large dynamic datasets of unlabeled images. In these cases, training or fine-tuning a model every time new images are added to the database is neither efficient nor scalable. Convolutional neural networks trained for image classification over large datasets have been proven effective feature extractors for image retrieval. The most successful approaches are based on encoding the activations of convolutional layers, as they convey the image spatial information. In this paper, we go beyond this spatial information and propose a local-aware encoding of convolutional features based on semantic information predicted in the target image. To this end, we obtain the most discriminative regions of an image using Class Activation Maps (CAMs). CAMs are based on the knowledge contained in the network and therefore, our approach, has the additional advantage of not requiring external information. In addition, we use CAMs to generate object proposals during an unsupervised re-ranking stage after a first fast search. Our experiments on two public available datasets for instance retrieval, Oxford5k and Paris6k, demonstrate the competitiveness of our approach outperforming the current state-of-the-art when using off-the-shelf models trained on ImageNet. The source code and model used in this paper are publicly available at http://imatge-upc.github.io/retrieval-2017-cam/.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision