z-logo
Premium
Lasso regression: estimation and shrinkage via the limit of Gibbs sampling
Author(s) -
Rajaratnam Bala,
Roberts Steven,
Sparks Doug,
Dalal Onkar
Publication year - 2016
Publication title -
journal of the royal statistical society: series b (statistical methodology)
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 6.523
H-Index - 137
eISSN - 1467-9868
pISSN - 1369-7412
DOI - 10.1111/rssb.12106
Subject(s) - multicollinearity , lasso (programming language) , estimator , elastic net regularization , algorithm , mathematics , gibbs sampling , computer science , bayesian probability , shrinkage estimator , linear regression , statistics , regression , bias of an estimator , minimum variance unbiased estimator , world wide web
Summary The application of the lasso is espoused in high dimensional settings where only a small number of the regression coefficients are believed to be non‐zero (i.e. the solution is sparse). Moreover, statistical properties of high dimensional lasso estimators are often proved under the assumption that the correlation between the predictors is bounded. In this vein, co‐ordinatewise methods, which are the most common means of computing the lasso solution, naturally work well in the presence of low‐to‐moderate multicollinearity. The computational speed of co‐ordinatewise algorithms, although excellent for sparse and low‐to‐moderate multicollinearity settings, degrades as sparsity decreases and multicollinearity increases. Though lack of sparsity and high multicollinearity can be quite common in contemporary applications, model selection is still a necessity in such settings. Motivated by the limitations of co‐ordinatewise algorithms in such ‘non‐sparse’ and ‘high multicollinearity’ settings, we propose the novel ‘deterministic Bayesian lasso’ algorithm for computing the lasso solution. This algorithm is developed by considering a limiting version of the Bayesian lasso. In contrast with co‐ordinatewise algorithms, the performance of the deterministic Bayesian lasso improves as sparsity decreases and multicollinearity increases. Importantly, in non‐sparse and high multicollinearity settings the algorithm proposed can offer substantial increases in computational speed over co‐ordinatewise algorithms. A rigorous theoretical analysis demonstrates that the deterministic Bayesian lasso algorithm converges to the lasso solution and it leads to a representation of the lasso estimator which shows how it achieves both l 1 ‐ and l 2 ‐types of shrinkage simultaneously. Connections between the deterministic Bayesian lasso and other algorithms are also provided. The benefits of the deterministic Bayesian lasso algorithm are then illustrated on simulated and real data.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here