Bias oriented unbiased data augmentation for cross-bias representation learning

The biased cues in the training data may build strong connections between specific targets and unexpected concepts, leading the learned representations could not be applied to real-world data that does not contain the same biased cues. To learn cross-bias representations which can generalize on unbi...

Full description

Saved in:
Bibliographic Details
Published inMultimedia systems Vol. 29; no. 2; pp. 725 - 738
Main Authors Li, Lei, Tang, Fan, Cao, Juan, Li, Xirong, Wang, Danding
Format Journal Article
LanguageEnglish
Published Berlin/Heidelberg Springer Berlin Heidelberg 01.04.2023
Springer Nature B.V
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The biased cues in the training data may build strong connections between specific targets and unexpected concepts, leading the learned representations could not be applied to real-world data that does not contain the same biased cues. To learn cross-bias representations which can generalize on unbiased datasets by only using biased data, researchers focus on reducing the influence of biased cues through unbiased sampling or augmentation on the basis of artificial experience. However, the distributions of biased cues in the dataset are neglected, which limits the performance of these methods. In this paper, we propose a bias oriented data augmentation to enhance the cross-bias generalization by enlarging “safety” and “unbiasedness” constraints in the training data without manual prior intervention. The safety constraint is proposed to maintain the class-specific information for augmentation while the unbiasedness constraint reduces the statistical correlation of bias information and class labels. Experiments under different biased proportions on four synthetic/real-world datasets show that the proposed approach could improve the performance of other SOTA debiasing approaches (colored MNIST: 0.35–26.14%, corrupted CIFAR10: 3.14–8.44%, BFFHQ: 1.50% and BAR: 1.72%).
ISSN:0942-4962
1432-1882
DOI:10.1007/s00530-022-01013-6