arXiv Analytics

Sign in

arXiv:1902.10755 [cs.LG]AbstractReferencesReviewsResources

Adversarial Attacks on Time Series

Fazle Karim, Somshubra Majumdar, Houshang Darabi

Published 2019-02-27Version 1

Time series classification models have been garnering significant importance in the research community. However, not much research has been done on generating adversarial samples for these models. These adversarial samples can become a security concern. In this paper, we propose utilizing an adversarial transformation network (ATN) on a distilled model to attack various time series classification models. The proposed attack on the classification model utilizes a distilled model as a surrogate that mimics the behavior of the attacked classical time series classification models. Our proposed methodology is applied onto 1-Nearest Neighbor Dynamic Time Warping (1-NN ) DTW, a Fully Connected Network and a Fully Convolutional Network (FCN), all of which are trained on 43 University of California Riverside (UCR) datasets. In this paper, we show both models were susceptible to attacks on all 43 datasets. To the best of our knowledge, such an attack on time series classification models has never been done before. Finally, we recommend future researchers that develop time series classification models to incorporating adversarial data samples into their training data sets to improve resilience on adversarial samples and to consider model robustness as an evaluative metric.

Related articles: Most relevant | Search more
arXiv:2004.00410 [cs.LG] (Published 2020-03-31)
Adversarial Attacks on Multivariate Time Series
arXiv:1802.06552 [cs.LG] (Published 2018-02-19)
Are Generative Classifiers More Robust to Adversarial Attacks?
arXiv:2002.03839 [cs.LG] (Published 2020-02-10)
Adversarial Attacks on Linear Contextual Bandits