Evaluation methodology for deep learning imputation models

There is growing interest in imputing missing data in tabular datasets using deep learning. Existing deep learning–based imputation models have been commonly evaluated using root mean square error (RMSE) as the predictive accuracy metric. In this article, we investigate the limitations of assessing...

Full description

Saved in:
Bibliographic Details
Published inExperimental biology and medicine (Maywood, N.J.) Vol. 247; no. 22; pp. 1972 - 1987
Main Authors Boursalie, Omar, Samavi, Reza, Doyle, Thomas E.
Format Journal Article
LanguageEnglish
Published London, England SAGE Publications 01.11.2022
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:There is growing interest in imputing missing data in tabular datasets using deep learning. Existing deep learning–based imputation models have been commonly evaluated using root mean square error (RMSE) as the predictive accuracy metric. In this article, we investigate the limitations of assessing deep learning–based imputation models by conducting a comparative analysis between RMSE and alternative metrics in the statistical literature including qualitative, predictive accuracy, statistical distance, and descriptive statistics. We design a new aggregated metric, called reconstruction loss (RL), to evaluate deep learning–based imputation models. We also develop and evaluate a novel imputation evaluation methodology based on RL. To minimize model and dataset biases, we use a regression imputation model and two different deep learning imputation models: denoising autoencoders and generative adversarial nets. We also use two tabular datasets from different industry sectors: health care and financial. Our results show that the proposed methodology is effective in evaluating multiple properties of the deep learning–based imputation model’s reconstruction performance.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:1535-3702
1535-3699
DOI:10.1177/15353702221121602