Steganographic universal adversarial perturbations

S. Ud Din, N. Akhtar, S. Younis, F. Shafait, A. Mansoor, M. Shafique

Research output: Contribution to journalArticlepeer-review

Abstract

We propose a steganography based technique to generate adversarial perturbations to fool deep models on any image. The proposed perturbations are computed in a transform domain where a single secret image embedded in any target image makes any deep model misclassify the target image with high probability. The attack resulting from our perturbation is ideal for black-box setting, as it does not require any information about the target model. Moreover, being a non-iterative technique, our perturbation estimation remains computationally efficient. The computed perturbations are also imperceptible to humans while they achieve high fooling ratios for the models trained on large-scale ImageNet dataset. We demonstrate successful fooling of ResNet-50, VGG-16, Inception-V3 and MobileNet-V2, achieving up to 89% fooling of these popular classification models.

Original languageEnglish (US)
Pages (from-to)146-152
Number of pages7
JournalPattern Recognition Letters
Volume135
DOIs
StatePublished - Jul 2020

Keywords

  • Adversarial attack
  • Deep neural networks
  • Steganography
  • Wavelet transform

ASJC Scopus subject areas

  • Software
  • Signal Processing
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Steganographic universal adversarial perturbations'. Together they form a unique fingerprint.

Cite this