The Bayesian bridge
We propose the Bayesian bridge estimator for regularized regression and classification. Two key mixture representations for the Bayesian bridge model are developed: a scale mixture of normal distributions with respect to an »-stable random variable; a mixture of Bartlett–Fejer kernels (or triangle d...
Saved in:
Published in | Journal of the Royal Statistical Society. Series B, Statistical methodology Vol. 76; no. 4; pp. 713 - 733 |
---|---|
Main Authors | , , |
Format | Journal Article |
Language | English |
Published |
Oxford
Blackwell Publishing Ltd
01.09.2014
John Wiley & Sons Ltd Oxford University Press |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We propose the Bayesian bridge estimator for regularized regression and classification. Two key mixture representations for the Bayesian bridge model are developed: a scale mixture of normal distributions with respect to an »-stable random variable; a mixture of Bartlett–Fejer kernels (or triangle densities) with respect to a two-component mixture of gamma random variables. Both lead to Markov chain Monte Carlo methods for posterior simulation, and these methods turn out to have complementary domains of maximum efficiency. The first representation is a well-known result due to West and is the better choice for collinear design matrices. The second representation is new and is more efficient for orthogonal problems, largely because it avoids the need to deal with exponentially tilted stable random variables. It also provides insight into the multimodality of the joint posterior distribution, which is a feature of the bridge model that is notably absent under ridge or lasso-type priors. We prove a theorem that extends this representation to a wider class of densities representable as scale mixtures of beta distributions, and we provide an explicit inversion formula for the mixing distribution. The connections with slice sampling and scale mixtures of normal distributions are explored. On the practical side, we find that the Bayesian bridge model outperforms its classical cousin in estimation and prediction across a variety of data sets, both simulated and real. We also show that the Markov chain Monte Carlo algorithm for fitting the bridge model exhibits excellent mixing properties, particularly for the global scale parameter. This makes for a favourable contrast with analogous Markov chain Monte Carlo algorithms for other sparse Bayesian models. All methods described in this paper are implemented in the R package BayesBridge. An extensive set of simulation results is provided in two on-line supplemental files. |
---|---|
Bibliography: | 'Benchmarking the two MCMC strategies for sampling the Bayesian bridge posterior distribution' and'An empirical study of mixing rates in parameter-expanded Gibbs samplers for sparse Bayesian regression models'. ArticleID:RSSB12042 istex:94C6BC2AC5884AC31C35517D520878C9F97FC94C ark:/67375/WNG-DND50R4L-J ObjectType-Article-2 SourceType-Scholarly Journals-1 ObjectType-Feature-1 content type line 23 |
ISSN: | 1369-7412 1467-9868 |
DOI: | 10.1111/rssb.12042 |