TY - JOUR
T1 - Perturbation diversity certificates robust generalization
AU - Qian, Zhuang
AU - Zhang, Shufei
AU - Huang, Kaizhu
AU - Wang, Qiufeng
AU - Yi, Xinping
AU - Gu, Bin
AU - Xiong, Huan
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2024/4
Y1 - 2024/4
N2 - Whilst adversarial training has been proven to be one most effective defending method against adversarial attacks for deep neural networks, it suffers from over-fitting on training adversarial data and thus may not guarantee the robust generalization. This may result from the fact that the conventional adversarial training methods generate adversarial perturbations usually in a supervised way so that the resulting adversarial examples are highly biased towards the decision boundary, leading to an inhomogeneous data distribution. To mitigate this limitation, we propose to generate adversarial examples from a perturbation diversity perspective. Specifically, the generated perturbed samples are not only adversarial but also diverse so as to certify robust generalization and significant robustness improvement through a homogeneous data distribution. We provide theoretical and empirical analysis, establishing a foundation to support the proposed method. As a major contribution, we prove that promoting perturbations diversity can lead to a better robust generalization bound. To verify our methods’ effectiveness, we conduct extensive experiments over different datasets (e.g., CIFAR-10, CIFAR-100, SVHN) with different adversarial attacks (e.g., PGD, CW). Experimental results show that our method outperforms other state-of-the-art (e.g., PGD and Feature Scattering) in robust generalization performance.
AB - Whilst adversarial training has been proven to be one most effective defending method against adversarial attacks for deep neural networks, it suffers from over-fitting on training adversarial data and thus may not guarantee the robust generalization. This may result from the fact that the conventional adversarial training methods generate adversarial perturbations usually in a supervised way so that the resulting adversarial examples are highly biased towards the decision boundary, leading to an inhomogeneous data distribution. To mitigate this limitation, we propose to generate adversarial examples from a perturbation diversity perspective. Specifically, the generated perturbed samples are not only adversarial but also diverse so as to certify robust generalization and significant robustness improvement through a homogeneous data distribution. We provide theoretical and empirical analysis, establishing a foundation to support the proposed method. As a major contribution, we prove that promoting perturbations diversity can lead to a better robust generalization bound. To verify our methods’ effectiveness, we conduct extensive experiments over different datasets (e.g., CIFAR-10, CIFAR-100, SVHN) with different adversarial attacks (e.g., PGD, CW). Experimental results show that our method outperforms other state-of-the-art (e.g., PGD and Feature Scattering) in robust generalization performance.
KW - Adversarial examples
KW - Adversarial robustness
KW - Robust generalization
UR - http://www.scopus.com/inward/record.url?scp=85182390363&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2024.106117
DO - 10.1016/j.neunet.2024.106117
M3 - Article
C2 - 38232423
AN - SCOPUS:85182390363
SN - 0893-6080
VL - 172
JO - Neural Networks
JF - Neural Networks
M1 - 106117
ER -