Understanding Adversarial Robustness from Feature Maps of Convolutional Layers

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

The adversarial robustness of a neural network mainly relies on two factors: model capacity and antiperturbation ability. In this article, we study the antiperturbation ability of the network from the feature maps of convolutional layers. Our theoretical analysis discovers that larger convolutional feature maps before average pooling can contribute to better resistance to perturbations, but the conclusion is not true for max pooling. It brings new inspiration to the design of robust neural networks and urges us to apply these findings to improve existing architectures. The proposed modifications are very simple and only require upsampling the inputs or slightly modifying the stride configurations of downsampling operators. We verify our approaches on several benchmark neural network architectures, including AlexNet, VGG, RestNet18, and PreActResNet18. Nontrivial improvements in terms of both natural accuracy and adversarial robustness can be achieved under various attack and defense mechanisms. The code is available at https://github.com/MTandHJ/rcm.

Original languageEnglish
Pages (from-to)4690-4702
Number of pages13
JournalIEEE Transactions on Neural Networks and Learning Systems
Volume36
Issue number3
DOIs
StatePublished - 2025

Keywords

  • Adversarial robustness
  • antiperturbation ability
  • convolutional layer
  • feature maps
  • pooling

Fingerprint

Dive into the research topics of 'Understanding Adversarial Robustness from Feature Maps of Convolutional Layers'. Together they form a unique fingerprint.

Cite this