Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network

October 01, 2018 ยท Entered Twilight ยท ๐Ÿ› International Conference on Learning Representations

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, LICENSE, README.md, acc_under_attack.py, acc_under_attack.sh, attacker, datasets, main_adv.py, main_adv_vi.py, main_plain.py, main_rse.py, main_vi.py, models, snr_density.py, snr_density.sh, test.py, train_adv.sh, train_adv_vi.sh, train_plain.sh, train_rse.sh, train_vi.sh, transfer_attack.py, utils

Authors Xuanqing Liu, Yao Li, Chongruo Wu, Cho-Jui Hsieh arXiv ID 1810.01279 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.CR, stat.ML Citations 182 Venue International Conference on Learning Representations Repository https://github.com/xuanqing94/BayesianDefense โญ 61 Last Checked 1 month ago
Abstract
We present a new algorithm to train a robust neural network against adversarial attacks. Our algorithm is motivated by the following two ideas. First, although recent work has demonstrated that fusing randomness can improve the robustness of neural networks (Liu 2017), we noticed that adding noise blindly to all the layers is not the optimal way to incorporate randomness. Instead, we model randomness under the framework of Bayesian Neural Network (BNN) to formally learn the posterior distribution of models in a scalable way. Second, we formulate the mini-max problem in BNN to learn the best model distribution under adversarial attacks, leading to an adversarial-trained Bayesian neural net. Experiment results demonstrate that the proposed algorithm achieves state-of-the-art performance under strong attacks. On CIFAR-10 with VGG network, our model leads to 14\% accuracy improvement compared with adversarial training (Madry 2017) and random self-ensemble (Liu 2017) under PGD attack with $0.035$ distortion, and the gap becomes even larger on a subset of ImageNet.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning