arXiv Analytics

Sign in

arXiv:1812.01804 [cs.LG]AbstractReferencesReviewsResources

Random Spiking and Systematic Evaluation of Defenses Against Adversarial Examples

Huangyi Ge, Sze Yiu Chau, Ninghui Li

Published 2018-12-05Version 1

Image classifiers often suffer from adversarial examples, which are generated by adding a small amount of noises to input images to trick classifiers into misclassification. Over the years, many defense mechanisms have been proposed, and different researchers have made seemingly contradictory claims on their effectiveness. We argue that such discrepancies are primarily due to inconsistent assumptions on the attacker's knowledge. To this end, we present an analysis of possible adversarial models, and propose an evaluation framework for comparing different defense mechanisms. As part of the framework, we introduced a more powerful and realistic adversary strategy. We propose a new defense mechanism called Random Spiking (RS), which generalizes dropout and introduces random noises in the training process in a controlled manner. With a carefully chosen placement, RS incurs negligible negative impact on prediction accuracy. Evaluations under our proposed framework suggest RS delivers better protection against adversarial examples than many existing schemes.

Related articles: Most relevant | Search more
arXiv:1911.05268 [cs.LG] (Published 2019-11-13)
Adversarial Examples in Modern Machine Learning: A Review
arXiv:1901.10861 [cs.LG] (Published 2019-01-30)
A Simple Explanation for the Existence of Adversarial Examples with Small Hamming Distance
arXiv:1906.07982 [cs.LG] (Published 2019-06-19)
A unified view on differential privacy and robustness to adversarial examples