Adversarially Robust Conformal Prediction

Asaf Gendler, Tsui-Wei Weng, Luca Daniel, Yaniv Romano

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Conformal prediction is a model-agnostic tool for constructing prediction sets that are valid under the common i.i.d. assumption, which has been applied to quantify the prediction uncertainty of deep net classifiers. In this paper, we generalize this framework to the case where adversaries exist during inference time, under which the i.i.d. assumption is grossly violated. By combining conformal prediction with randomized smoothing, our proposed method forms a prediction set with finite-sample coverage guarantee that holds for any data distribution with 2-norm bounded adversarial noise, generated by any adversarial attack algorithm. The core idea is to bound the Lipschitz constant of the non-conformity score by smoothing it with Gaussian noise and leverage this knowledge to account for the effect of the unknown adversarial perturbation. We demonstrate the necessity of our method in the adversarial setting and the validity of our theoretical guarantee on three widely used benchmark data sets: CIFAR10, CIFAR100, and ImageNet.
Original languageAmerican English
Title of host publicationICLR
StatePublished - 2022
EventInternational Conference on Learning Representations -
Duration: 25 Apr 2022 → …

Conference

ConferenceInternational Conference on Learning Representations
Period25/04/22 → …

Cite this