Abstract:
Transfer-based attacks against black-box neural network models have received increasing attention because they are more realistic scenarios, but how to produce highly tra...Show MoreMetadata
Abstract:
Transfer-based attacks against black-box neural network models have received increasing attention because they are more realistic scenarios, but how to produce highly transferable adversarial examples on the surrogate model becomes critical. In this work, we find that if the attack direction of the original example is controlled from the beginning, the produced adversarial examples will be more transferable. Specifically, we propose the Output Direction Controller (ODC) to initialize the example direction so that the example starts off with a deviation from the true direction or toward the target direction. ODC is a simple and extensible component that can be combined with various transfer-based attack methods and significantly improve the transferability of the adversarial examples. On the ImageNet dataset, we optimize the baseline method by ODC to improve the success rate of untargeted attacks by an average of 11.79% and targeted attacks by an average of 3.38%. Code is available at https://github.com/yangrongbo/ODC.
Date of Conference: 01-04 October 2023
Date Added to IEEE Xplore: 29 January 2024
ISBN Information:
ISSN Information:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Neural Network ,
- Artificial Neural Network ,
- Alternative Models ,
- Baseline Methods ,
- Attack Rate ,
- ImageNet Dataset ,
- Attack Target ,
- Adversarial Examples ,
- Attack Methods ,
- Attack Success Rate ,
- Deep Neural Network ,
- Direction Of Change ,
- Ensemble Model ,
- Layer Model ,
- Clear Image ,
- Target Model ,
- Cross-entropy Loss Function ,
- Random Perturbations ,
- Adversarial Attacks ,
- Output Space ,
- Black-box Attacks ,
- Fast Gradient Sign Method ,
- White-box Attack ,
- Gradient Ascent ,
- Highest Success Rate ,
- Real Label
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Neural Network ,
- Artificial Neural Network ,
- Alternative Models ,
- Baseline Methods ,
- Attack Rate ,
- ImageNet Dataset ,
- Attack Target ,
- Adversarial Examples ,
- Attack Methods ,
- Attack Success Rate ,
- Deep Neural Network ,
- Direction Of Change ,
- Ensemble Model ,
- Layer Model ,
- Clear Image ,
- Target Model ,
- Cross-entropy Loss Function ,
- Random Perturbations ,
- Adversarial Attacks ,
- Output Space ,
- Black-box Attacks ,
- Fast Gradient Sign Method ,
- White-box Attack ,
- Gradient Ascent ,
- Highest Success Rate ,
- Real Label
- Author Keywords