Strengthening transferability of adversarial examples by adaptive inertia and amplitude spectrum dropout.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

Deep neural networks are sensitive to adversarial examples and would produce wrong results with high confidence. However, most existing attack methods exhibit weak transferability, especially for adversarially trained models and defense models. In this paper, two methods are proposed to generate highly transferable adversarial examples, namely Adaptive Inertia Iterative Fast Gradient Sign Method (AdaI-FGSM) and Amplitude Spectrum Dropout Method (ASDM). Specifically, AdaI-FGSM aims to integrate adaptive inertia into the gradient-based attack, and leverage the looking ahead property to search for a flatter maximum, which is essential to strengthen the transferability of adversarial examples. By introducing a loss-preserving transformation in the frequency domain, the proposed ASDM with the dropout invariance property can craft the copies of input images to overcome the poor generalization on the surrogate models. Furthermore, AdaI-FGSM and ASDM can be naturally integrated as an efficient gradient-based attack method to yield more transferable adversarial examples. Extensive experimental results on the ImageNet-compatible dataset demonstrate that higher transferability is achieved by our method than some advanced gradient-based attacks.

Authors

  • Huanhuan Li
    School of Food and Biological Engineering, Jiangsu University, Zhenjiang 212013, PR China.
  • Wenbo Yu
    Beijing Advanced Innovation Center for Food Nutrition and Human Health, College of Veterinary Medicine, China Agricultural University, Beijing Key Laboratory of Detection Technology for Animal-Derived Food Safety, And Beijing Laboratory for Food Quality and Safety, Beijing, 100193, People's Republic of China.
  • He Huang