Robust Wasserstein profile inference and applications to machine learning
We show that several machine learning estimators, including square-root least absolute shrinkage and selection and regularized logistic regression, can be represented as solutions to distributionally robust optimization problems. The associated uncertainty regions are based on suitably defined Wasse...
Saved in:
Published in | Journal of applied probability Vol. 56; no. 3; pp. 830 - 857 |
---|---|
Main Authors | , , |
Format | Journal Article |
Language | English |
Published |
Cambridge, UK
Cambridge University Press
01.09.2019
Applied Probability Trust |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We show that several machine learning estimators, including square-root least absolute shrinkage and selection and regularized logistic regression, can be represented as solutions to distributionally robust optimization problems. The associated uncertainty regions are based on suitably defined Wasserstein distances. Hence, our representations allow us to view regularization as a result of introducing an artificial adversary that perturbs the empirical distribution to account for out-of-sample effects in loss estimation. In addition, we introduce RWPI (robust Wasserstein profile inference), a novel inference methodology which extends the use of methods inspired by empirical likelihood to the setting of optimal transport costs (of which Wasserstein distances are a particular case). We use RWPI to show how to optimally select the size of uncertainty regions, and as a consequence we are able to choose regularization parameters for these machine learning estimators without the use of cross validation. Numerical experiments are also given to validate our theoretical findings. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 |
ISSN: | 0021-9002 1475-6072 |
DOI: | 10.1017/jpr.2019.49 |