DelugeNets: Deep Networks with Efficient and Flexible Cross-layer Information Inflows

November 17, 2016 ยท Entered Twilight ยท ๐Ÿ› 2017 IEEE International Conference on Computer Vision Workshops (ICCVW)

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 8.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: CrossLayerDepthwiseConvolution.lua, LICENSE, README.md, delugenet.lua

Authors Jason Kuen, Xiangfei Kong, Gang Wang, Yap-Peng Tan arXiv ID 1611.05552 Category cs.CV: Computer Vision Cross-listed cs.LG, cs.NE Citations 16 Venue 2017 IEEE International Conference on Computer Vision Workshops (ICCVW) Repository https://github.com/xternalz/DelugeNets โญ 26 Last Checked 1 month ago
Abstract
Deluge Networks (DelugeNets) are deep neural networks which efficiently facilitate massive cross-layer information inflows from preceding layers to succeeding layers. The connections between layers in DelugeNets are established through cross-layer depthwise convolutional layers with learnable filters, acting as a flexible yet efficient selection mechanism. DelugeNets can propagate information across many layers with greater flexibility and utilize network parameters more effectively compared to ResNets, whilst being more efficient than DenseNets. Remarkably, a DelugeNet model with just model complexity of 4.31 GigaFLOPs and 20.2M network parameters, achieve classification errors of 3.76% and 19.02% on CIFAR-10 and CIFAR-100 dataset respectively. Moreover, DelugeNet-122 performs competitively to ResNet-200 on ImageNet dataset, despite costing merely half of the computations needed by the latter.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision