Sparse dimension reduction based on energy and ball statistics
Two new methods for sparse dimension reduction are introduced, based on martingale difference divergence and ball covariance, respectively. These methods can be utilized straightforwardly as sufficient dimension reduction (SDR) techniques to estimate a sufficient dimension reduced subspace, which co...
Saved in:
Published in | Advances in data analysis and classification Vol. 16; no. 4; pp. 951 - 975 |
---|---|
Main Authors | , , |
Format | Journal Article |
Language | English |
Published |
Berlin/Heidelberg
Springer Berlin Heidelberg
01.12.2022
Springer Nature B.V |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Two new methods for sparse dimension reduction are introduced, based on martingale difference divergence and ball covariance, respectively. These methods can be utilized straightforwardly as sufficient dimension reduction (SDR) techniques to estimate a sufficient dimension reduced subspace, which contains all information sufficient to explain a dependent variable. Moreover, owing to their sparsity, they intrinsically perform sufficient variable selection (SVS) and present two attractive new approaches to variable selection in a context of nonlinear dependencies that require few model assumptions. The two new methods are compared to a similar existing approach for SDR and SVS based on distance covariance, as well as to classical and robust sparse partial least squares. A simulation study shows that each of the new estimators can achieve correct variable selection in highly nonlinear contexts, yet are sensitive to outliers and computationally intensive. The study sheds light on the subtle differences between the methods. Two examples illustrate how they can be applied in practice, with a slight preference for the option based on martingale difference divergence in a bioinformatics example. |
---|---|
ISSN: | 1862-5347 1862-5355 |
DOI: | 10.1007/s11634-021-00470-7 |