Random Smoothing Might be Unable to Certify $\ell_\infty$ Robustness for High-Dimensional Images

February 10, 2020 ยท Entered Twilight ยท ๐Ÿ› Journal of machine learning research

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE, README.md, checkpoints, code, results

Authors Avrim Blum, Travis Dick, Naren Manoj, Hongyang Zhang arXiv ID 2002.03517 Category cs.LG: Machine Learning Cross-listed cs.CR, stat.ML Citations 83 Venue Journal of machine learning research Repository https://github.com/hongyanz/TRADES-smoothing โญ 14 Last Checked 1 month ago
Abstract
We show a hardness result for random smoothing to achieve certified adversarial robustness against attacks in the $\ell_p$ ball of radius $ฮต$ when $p>2$. Although random smoothing has been well understood for the $\ell_2$ case using the Gaussian distribution, much remains unknown concerning the existence of a noise distribution that works for the case of $p>2$. This has been posed as an open problem by Cohen et al. (2019) and includes many significant paradigms such as the $\ell_\infty$ threat model. In this work, we show that any noise distribution $\mathcal{D}$ over $\mathbb{R}^d$ that provides $\ell_p$ robustness for all base classifiers with $p>2$ must satisfy $\mathbb{E}ฮท_i^2=ฮฉ(d^{1-2/p}ฮต^2(1-ฮด)/ฮด^2)$ for 99% of the features (pixels) of vector $ฮท\sim\mathcal{D}$, where $ฮต$ is the robust radius and $ฮด$ is the score gap between the highest-scored class and the runner-up. Therefore, for high-dimensional images with pixel values bounded in $[0,255]$, the required noise will eventually dominate the useful information in the images, leading to trivial smoothed classifiers.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning