Towards Practical Control of Singular Values of Convolutional Layers

November 24, 2022 ยท Entered Twilight ยท ๐Ÿ› Neural Information Processing Systems

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, Dockerfile, LICENSE, README.md, __init__.py, doc, experiments.sh, requirements.txt, requirements_reproducibility.txt, src

Authors Alexandra Senderovich, Ekaterina Bulatova, Anton Obukhov, Maxim Rakhuba arXiv ID 2211.13771 Category cs.LG: Machine Learning Cross-listed cs.CV Citations 12 Venue Neural Information Processing Systems Repository https://github.com/WhiteTeaDragon/practical_svd_conv โญ 9 Last Checked 1 month ago
Abstract
In general, convolutional neural networks (CNNs) are easy to train, but their essential properties, such as generalization error and adversarial robustness, are hard to control. Recent research demonstrated that singular values of convolutional layers significantly affect such elusive properties and offered several methods for controlling them. Nevertheless, these methods present an intractable computational challenge or resort to coarse approximations. In this paper, we offer a principled approach to alleviating constraints of the prior art at the expense of an insignificant reduction in layer expressivity. Our method is based on the tensor-train decomposition; it retains control over the actual singular values of convolutional mappings while providing structurally sparse and hardware-friendly representation. We demonstrate the improved properties of modern CNNs with our method and analyze its impact on the model performance, calibration, and adversarial robustness. The source code is available at: https://github.com/WhiteTeaDragon/practical_svd_conv
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning