Debiased-CAM to mitigate image perturbations with faithful visual explanations of machine learning
Model explanations such as saliency maps can improve user trust in AI by highlighting important features for a prediction. However, these become distorted and misleading when explaining predictions of images that are subject to systematic error (bias) by perturbations and corruptions. Furthermore, t...
Saved in:
Main Authors | , , |
---|---|
Format | Journal Article |
Language | English |
Published |
10.12.2020
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.2012.05567 |
Cover
Loading…
Summary: | Model explanations such as saliency maps can improve user trust in AI by
highlighting important features for a prediction. However, these become
distorted and misleading when explaining predictions of images that are subject
to systematic error (bias) by perturbations and corruptions. Furthermore, the
distortions persist despite model fine-tuning on images biased by different
factors (blur, color temperature, day/night). We present Debiased-CAM to
recover explanation faithfulness across various bias types and levels by
training a multi-input, multi-task model with auxiliary tasks for explanation
and bias level predictions. In simulation studies, the approach not only
enhanced prediction accuracy, but also generated highly faithful explanations
about these predictions as if the images were unbiased. In user studies,
debiased explanations improved user task performance, perceived truthfulness
and perceived helpfulness. Debiased training can provide a versatile platform
for robust performance and explanation faithfulness for a wide range of
applications with data biases. |
---|---|
DOI: | 10.48550/arxiv.2012.05567 |