Adversarial Vulnerability of Neural Networks Increases With Input Dimension
Carl-Johann Simon-Gabriel
and
Yann Ollivier
and
Léon Bottou
and
Bernhard Schölkopf
and
David Lopez-Paz
arXiv e-Print archive - 2018 via Local arXiv
Keywords:
stat.ML, cs.CV, cs.LG, 68T45, I.2.6
First published: 2018/02/05 (6 years ago) Abstract: Over the past four years, neural networks have proven vulnerable to
adversarial images: targeted but imperceptible image perturbations lead to
drastically different predictions. We show that adversarial vulnerability
increases with the gradients of the training objective when seen as a function
of the inputs. For most current network architectures, we prove that the
$\ell_1$-norm of these gradients grows as the square root of the input-size.
These nets therefore become increasingly vulnerable with growing image size.
Over the course of our analysis we rediscover and generalize
double-backpropagation, a technique that penalizes large gradients in the loss
surface to reduce adversarial vulnerability and increase generalization
performance. We show that this regularization-scheme is equivalent at first
order to training with adversarial noise. Our proofs rely on the network's
weight-distribution at initialization, but extensive experiments confirm all
conclusions after training.