Robust Training with Feature-Based Adversarial Example

Xuanming Fu, Zhengfeng Yang, Hao Xue, Jianlin Wang, Zhenbing Zeng

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Adversarial training is an efficacious defense approach to protect classification model against adversarial attacks. In this paper, we reveal that a significant difference exists between the feature map of the original sample and that of its corresponding adversarial version. Based on this main insight, we propose a novel robust training on feature-based adversarial examples approach called FPAT, where training examples are generated by maximizing the loss function between the clean and the adversarial feature maps. We show via extensive experiments on MNIST, SVHN and CIFAR-10, that our proposed method is as effective as the state-of-the-art robust training methods. Especially, when the adversarial perturbation is of a large radius or the number of adversarial steps of training samples is small, FPAT achieves leading robustness.

Original languageEnglish
Title of host publication2022 26th International Conference on Pattern Recognition, ICPR 2022
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages2957-2963
Number of pages7
ISBN (Electronic)9781665490627
DOIs
StatePublished - 2022
Event26th International Conference on Pattern Recognition, ICPR 2022 - Montreal, Canada
Duration: 21 Aug 202225 Aug 2022

Publication series

NameProceedings - International Conference on Pattern Recognition
Volume2022-August
ISSN (Print)1051-4651

Conference

Conference26th International Conference on Pattern Recognition, ICPR 2022
Country/TerritoryCanada
CityMontreal
Period21/08/2225/08/22

Fingerprint

Dive into the research topics of 'Robust Training with Feature-Based Adversarial Example'. Together they form a unique fingerprint.

Cite this