Adaptive iterative attack towards explainable adversarial robustness

Yucheng Shi, Yahong Han*, Quanxin Zhang, Xiaohui Kuang

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

46 引用 (Scopus)

摘要

Image classifiers based on deep neural networks show severe vulnerability when facing adversarial examples crafted on purpose. Designing more effective and efficient adversarial attacks is attracting considerable interest due to its potential contribution to interpretability of deep learning and validation of neural networks’ robustness. However, current iterative attacks use a fixed step size for each noise-adding step, making further investigation into the effect of variable step size on model robustness ripe for exploration. We prove that if the upper bound of noise added to the original image is fixed, the attack effect can be improved if the step size is positively correlated with the gradient obtained at each step by querying the target model. In this paper, we propose Ada-FGSM (Adaptive FGSM), a new iterative attack that adaptively allocates step size of noises according to gradient information at each step. Improvement of success rate and accuracy decrease measured on ImageNet with multiple models emphasizes the validity of our method. We analyze the process of iterative attack by visualizing their trajectory and gradient contour, and further explain the vulnerability of deep neural networks to variable step size adversarial examples.

源语言英语
文章编号107309
期刊Pattern Recognition
105
DOI
出版状态已出版 - 9月 2020

指纹

探究 'Adaptive iterative attack towards explainable adversarial robustness' 的科研主题。它们共同构成独一无二的指纹。

引用此