Semantic Perturbations with Normalizing Flows for Improved Generalization

Data augmentation is a widely adopted technique for avoiding overfitting when training deep neural networks. However, this approach requires domain-specific knowledge and is often limited to a fixed set of hard-coded transformations. Recently, several works proposed to use generative models for gene...

Full description

Saved in:
Bibliographic Details
Published in2021 IEEE/CVF International Conference on Computer Vision (ICCV) pp. 6599 - 6609
Main Authors Yuksel, Oguz Kaan, Stich, Sebastian U., Jaggi, Martin, Chavdarova, Tatjana
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.10.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Data augmentation is a widely adopted technique for avoiding overfitting when training deep neural networks. However, this approach requires domain-specific knowledge and is often limited to a fixed set of hard-coded transformations. Recently, several works proposed to use generative models for generating semantically meaningful perturbations to train a classifier. However, because accurate encoding and decoding are critical, these methods, which use architectures that approximate the latent-variable inference, remained limited to pilot studies on small datasets.Exploiting the exactly reversible encoder-decoder structure of normalizing flows, we perform on-manifold perturbations in the latent space to define fully unsupervised data augmentations. We demonstrate that such perturbations match the performance of advanced data augmentation techniques-reaching 96.6% test accuracy for CIFAR10 using ResNet-18 and outperform existing methods, particularly in low data regimes-yielding 10-25% relative improvement of test accuracy from classical training. We find that our latent adversarial perturbations adaptive to the classifier throughout its training are most effective, yielding the first test accuracy improvement results on real-world datasets-CIFAR-10/100-via latent-space perturbations.
ISSN:2380-7504
DOI:10.1109/ICCV48922.2021.00655