Towards Robust Neural Networks via Random Self-ensemble Towards Robust Neural Networks via Random Self-ensemble
Paper summary Liu et al. propose randomizing neural networks, implicitly learning an ensemble of models, to defend against adversarial attacks. In particular, they introduce Gaussian noise layers before regular convolutional layers. The noise can be seen as additional parameter of the model. During training, noise is randomly added. During testing, the model is evaluated on a single testing input using multiple random noise vectors; this essentially corresponds to an ensemble of different models (parameterized by the different noise vectors). Mathemtically, the authors provide two interesting interpretations. First, they argue that training essentially minimizes an upper bound of the (noisy) inference loss. Second, they show that their approach is equivalent to Lipschitz regularization [1]. [1] M. Hein, M. Andriushchenko. Formal guarantees on the robustness of a classifier against adversarial manipulation. ArXiv:1705.08475, 2017. Also view this summary at [](
Towards Robust Neural Networks via Random Self-ensemble
Xuanqing Liu and Minhao Cheng and Huan Zhang and Cho-Jui Hsieh
arXiv e-Print archive - 2017 via Local arXiv
Keywords: cs.LG, cs.CR, stat.ML


Summary by David Stutz 1 month ago
Your comment: allows researchers to publish paper summaries that are voted on and ranked!

Sponsored by: and