arXiv Analytics

Sign in

arXiv:1911.08696 [cs.LG]AbstractReferencesReviewsResources

Where is the Bottleneck of Adversarial Learning with Unlabeled Data?

Jingfeng Zhang, Bo Han, Gang Niu, Tongliang Liu, Masashi Sugiyama

Published 2019-11-20Version 1

Deep neural networks (DNNs) are incredibly brittle due to adversarial examples. To robustify DNNs, adversarial training was proposed, which requires large-scale but well-labeled data. However, it is quite expensive to annotate large-scale data well. To compensate for this shortage, several seminal works are utilizing large-scale unlabeled data. In this paper, we observe that seminal works do not perform well, since the quality of pseudo labels on unlabeled data is quite poor, especially when the amount of unlabeled data is significantly larger than that of labeled data. We believe that the quality of pseudo labels is the bottleneck of adversarial learning with unlabeled data. To tackle this bottleneck, we leverage deep co-training, which trains two deep networks and encourages two networks diverged by exploiting peer's adversarial examples. Based on deep co-training, we propose robust co-training (RCT) for adversarial learning with unlabeled data. We conduct comprehensive experiments on CIFAR-10 and SVHN datasets. Empirical results demonstrate that our RCT can significantly outperform baselines (e.g., robust self-training (RST)) in both standard test accuracy and robust test accuracy w.r.t. different datasets, different network structures, and different types of adversarial training.

Related articles: Most relevant | Search more
arXiv:1906.00555 [cs.LG] (Published 2019-06-03)
Adversarially Robust Generalization Just Requires More Unlabeled Data
arXiv:1905.11866 [cs.LG] (Published 2019-05-28)
When can unlabeled data improve the learning rate?
arXiv:1904.11717 [cs.LG] (Published 2019-04-26)
Classification from Pairwise Similarities/Dissimilarities and Unlabeled Data via Empirical Risk Minimization