This repository contains code for the implementation of our NeurIPS 2022 paper Toward Efficient Robust Training against Union of Lp Threat Models.
Recording for the NeurIPS Virtual Conference: [video]
In-person Presentation media: [poster]
To summarize, we make the following contributions in this paper:
- We demonstrate the first successful single-step robust training procedure, NCAT-$\ell_1$, to
achieve
$\ell_1$ robustness by using a curriculum schedule with Nuclear Norm based training. - We extend this approach to propose a training procedure NCAT, that yields SOTA-like robust
accuracy under the union of multiple
$\ell_p$ threat models, while requiring only a single-step attack budget per minibatch. - We further demonstrate that the proposed defense can scale-up to high-capacity networks and large-scale datasets such as ImageNet-100. Additionally, NCAT trained models generalize to unseen threat models, achieving near-SOTA robustness even on Perceptual Projected Gradient Descent (PPGD), which comprises one of the strongest attacks known to date.
In this work, we develop an efficient adversarial training procedure, NCAT, to train networks that are robust against a union of
- Key Observation: Setting up a curriculum for adversarial perturbations greatly improves overall stability, especially in sensitive
$\ell_1$ based training - However, RFGSM-AT with curriculum only leads to a delay in catastrophic failure; R-FGSM adversaries are not quite suitable for robust training
Catastrophic Overfitting in
NCAT: While pointwise loss minimization is susceptible to catastrophic failure, enforcing local function smoothness with a curriculum schedule enables the first successful demonstration of single-step training against
Trained model checkpoints can be found here.
Here we present a brief summary of results on the CIFAR-10 dataset obtained using the ResNet-18 architecture. For more details, please refer to the main paper. Robust evaluations are presented under the constraint sets given by the
@inproceedings{
sriramanan2022toward,
title={Toward Efficient Robust Training against Union of \${\textbackslash}ell\_p\$ Threat Models},
author={Gaurang Sriramanan and Maharshi Gor and Soheil Feizi},
booktitle={Advances in Neural Information Processing Systems},
editor={Alice H. Oh and Alekh Agarwal and Danielle Belgrave and Kyunghyun Cho},
year={2022},
url={https://openreview.net/forum?id=6qdUJblMHqy}
}