Defense against adversarial attacks by low-level image transformations

  • Zhaoxia Yin
  • , Hua Wang
  • , Jie Wang
  • , Jin Tang
  • , Wenzhong Wang*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

21 Scopus citations

Abstract

Deep neural networks (DNNs) are vulnerable to adversarial examples, which can fool classifiers by maliciously adding imperceptible perturbations to the original input. Currently, a large number of research on defending adversarial examples pay little attention to the real-world applications, either with high computational complexity or poor defensive effects. Motivated by this observation, we develop an efficient preprocessing module to defend adversarial attacks. Specifically, before an adversarial example is fed into the model, we perform two low-level image transformations, WebP compression and flip operation, on the picture. Then we can get a de-perturbed sample that can be correctly classified by DNNs. WebP compression is utilized to remove the small adversarial noises. Due to the introduction of loop filtering, there will be no square effect like JPEG compression, so the visual quality of the denoised image is higher. And flip operation, which flips the image once along one side of the image, destroys the specific structure of adversarial perturbations. By taking class activation mapping to localize the discriminative image regions, we show that flipping image may mitigate adversarial effects. Extensive experiments demonstrate that the proposed scheme outperforms the state-of-the-art defense methods. It can effectively defend adversarial attacks while ensuring only slight accuracy drops on normal images.

Original languageEnglish
Pages (from-to)1453-1466
Number of pages14
JournalInternational Journal of Intelligent Systems
Volume35
Issue number10
DOIs
StatePublished - 1 Oct 2020
Externally publishedYes

Keywords

  • WebP compression
  • adversarial examples
  • deep neural networks
  • flip operation
  • image transformations

Fingerprint

Dive into the research topics of 'Defense against adversarial attacks by low-level image transformations'. Together they form a unique fingerprint.

Cite this