arXiv Analytics

Sign in

arXiv:1802.06552 [cs.LG]AbstractReferencesReviewsResources

Are Generative Classifiers More Robust to Adversarial Attacks?

Yingzhen Li

Published 2018-02-19Version 1

There is a rising interest in studying the robustness of deep neural network classifiers against adversaries, with both advanced attack and defence techniques being actively developed. However, most recent work focuses on discriminative classifiers which only models the conditional distribution of the labels given the inputs. In this abstract we propose deep Bayes classifier that improves the classical naive Bayes with conditional deep generative models, and verifies its robustness against a number of existing attacks. We further developed a detection method for adversarial examples based on conditional deep generative models. Our initial results on MNIST suggest that deep Bayes classifiers might be more robust when compared with deep discriminative classifiers, and the proposed detection method achieves high detection rates against two commonly used attacks.

Comments: Submitted to ICLR 2018 workshop track, preliminary work, feedback welcome
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:2002.03839 [cs.LG] (Published 2020-02-10)
Adversarial Attacks on Linear Contextual Bandits
arXiv:1902.10755 [cs.LG] (Published 2019-02-27)
Adversarial Attacks on Time Series
arXiv:1811.06492 [cs.LG] (Published 2018-11-15)
Mathematical Analysis of Adversarial Attacks