Localized Uncertainty Attacks

The susceptibility of deep learning models to adversarial perturbations has stirred renewed attention in adversarial examples resulting in a number of attacks. However, most of these attacks fail to encompass a large spectrum of adversarial perturbations that are imperceptible to humans. In this pap...

Full description

Saved in:
Bibliographic Details
Published inarXiv.org
Main Authors Ousmane Amadou Dia, Karaletsos, Theofanis, Caner Hazirbas, Cristian Canton Ferrer, Ilknur Kaynar Kabul, Meijer, Erik
Format Paper
LanguageEnglish
Published Ithaca Cornell University Library, arXiv.org 17.06.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The susceptibility of deep learning models to adversarial perturbations has stirred renewed attention in adversarial examples resulting in a number of attacks. However, most of these attacks fail to encompass a large spectrum of adversarial perturbations that are imperceptible to humans. In this paper, we present localized uncertainty attacks, a novel class of threat models against deterministic and stochastic classifiers. Under this threat model, we create adversarial examples by perturbing only regions in the inputs where a classifier is uncertain. To find such regions, we utilize the predictive uncertainty of the classifier when the classifier is stochastic or, we learn a surrogate model to amortize the uncertainty when it is deterministic. Unlike \(\ell_p\) ball or functional attacks which perturb inputs indiscriminately, our targeted changes can be less perceptible. When considered under our threat model, these attacks still produce strong adversarial examples; with the examples retaining a greater degree of similarity with the inputs.
ISSN:2331-8422