arXiv Analytics

Sign in

arXiv:2003.06566 [cs.LG]AbstractReferencesReviewsResources

VarMixup: Exploiting the Latent Space for Robust Training and Inference

Puneet Mangla, Vedant Singh, Shreyas Jayant Havaldar, Vineeth N Balasubramanian

Published 2020-03-14Version 1

The vulnerability of Deep Neural Networks (DNNs) to adversarial attacks has led to the development of many defense approaches. Among them, Adversarial Training (AT) is a popular and widely used approach for training adversarially robust models. Mixup Training (MT), a recent popular training algorithm, improves the generalization performance of models by introducing globally linear behavior in between training examples. Although still in its early phase, we observe a shift in trend of exploiting Mixup from perspectives of generalisation to that of adversarial robustness. It has been shown that the Mixup trained models improves the robustness of models but only passively. A recent approach, Mixup Inference (MI), proposes an inference principle for Mixup trained models to counter adversarial examples at inference time by mixing the input with other random clean samples. In this work, we propose a new approach - \textit{VarMixup (Variational Mixup)} - to better sample mixup images by using the latent manifold underlying the data. Our experiments on CIFAR-10, CIFAR-100, SVHN and Tiny-Imagenet demonstrate that \textit{VarMixup} beats state-of-the-art AT techniques without training the model adversarially. Additionally, we also conduct ablations that show that models trained on \textit{VarMixup} samples are also robust to various input corruptions/perturbations, have low calibration error and are transferable.

Comments: 21 pages, 7 figures, 5 Tables
Categories: cs.LG, cs.AI, cs.CV
Related articles: Most relevant | Search more
arXiv:1710.02338 [cs.LG] (Published 2017-10-06)
Projection Based Weight Normalization for Deep Neural Networks
arXiv:1605.05359 [cs.LG] (Published 2016-05-17)
Hierarchical Reinforcement Learning using Spatio-Temporal Abstractions and Deep Neural Networks
arXiv:1511.05497 [cs.LG] (Published 2015-11-17)
Learning the Architecture of Deep Neural Networks