{ "id": "2408.12670", "version": "v1", "published": "2024-08-22T18:24:08.000Z", "updated": "2024-08-22T18:24:08.000Z", "title": "Leveraging Information Consistency in Frequency and Spatial Domain for Adversarial Attacks", "authors": [ "Zhibo Jin", "Jiayu Zhang", "Zhiyu Zhu", "Xinyi Wang", "Yiyun Huang", "Huaming Chen" ], "comment": "Accepted by PRICAI 2024", "categories": [ "cs.LG", "cs.AI" ], "abstract": "Adversarial examples are a key method to exploit deep neural networks. Using gradient information, such examples can be generated in an efficient way without altering the victim model. Recent frequency domain transformation has further enhanced the transferability of such adversarial examples, such as spectrum simulation attack. In this work, we investigate the effectiveness of frequency domain-based attacks, aligning with similar findings in the spatial domain. Furthermore, such consistency between the frequency and spatial domains provides insights into how gradient-based adversarial attacks induce perturbations across different domains, which is yet to be explored. Hence, we propose a simple, effective, and scalable gradient-based adversarial attack algorithm leveraging the information consistency in both frequency and spatial domains. We evaluate the algorithm for its effectiveness against different models. Extensive experiments demonstrate that our algorithm achieves state-of-the-art results compared to other gradient-based algorithms. Our code is available at: https://github.com/LMBTough/FSA.", "revisions": [ { "version": "v1", "updated": "2024-08-22T18:24:08.000Z" } ], "analyses": { "keywords": [ "spatial domain", "leveraging information consistency", "adversarial attacks induce perturbations", "gradient-based adversarial attack algorithm", "achieves state-of-the-art results" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }