Deep Neural Network Compression with Single and Multiple Level Quantization

March 06, 2018 ยท Entered Twilight ยท ๐Ÿ› AAAI Conference on Artificial Intelligence

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: CMakeLists.txt, CONTRIBUTING.md, CONTRIBUTORS.md, INSTALL.md, LICENSE, Makefile, Makefile.config, Makefile.config.example, README.md, caffe.cloc, cmake, data, docker, docs, examples, glmdb.sh, include, matlab, models, python, scripts, src, tools

Authors Yuhui Xu, Yongzhuang Wang, Aojun Zhou, Weiyao Lin, Hongkai Xiong arXiv ID 1803.03289 Category cs.LG: Machine Learning Cross-listed cs.AI, stat.ML Citations 122 Venue AAAI Conference on Artificial Intelligence Repository https://github.com/yuhuixu1993/SLQ โญ 19 Last Checked 1 month ago
Abstract
Network quantization is an effective solution to compress deep neural networks for practical usage. Existing network quantization methods cannot sufficiently exploit the depth information to generate low-bit compressed network. In this paper, we propose two novel network quantization approaches, single-level network quantization (SLQ) for high-bit quantization and multi-level network quantization (MLQ) for extremely low-bit quantization (ternary).We are the first to consider the network quantization from both width and depth level. In the width level, parameters are divided into two parts: one for quantization and the other for re-training to eliminate the quantization loss. SLQ leverages the distribution of the parameters to improve the width level. In the depth level, we introduce incremental layer compensation to quantize layers iteratively which decreases the quantization loss in each iteration. The proposed approaches are validated with extensive experiments based on the state-of-the-art neural networks including AlexNet, VGG-16, GoogleNet and ResNet-18. Both SLQ and MLQ achieve impressive results.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning