Least Angle Regression
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, least-angle regression (LARS) is an algorithm for fitting
linear regression In statistics, linear regression is a linear approach for modelling the relationship between a scalar response and one or more explanatory variables (also known as dependent and independent variables). The case of one explanatory variable is call ...
models to high-dimensional data, developed by
Bradley Efron Bradley Efron (; born May 24, 1938) is an American statistician. Efron has been president of the American Statistical Association (2004) and of the Institute of Mathematical Statistics (1987–1988).Cochran, J. (1 September 2015), "ASA Lead ...
,
Trevor Hastie Trevor John Hastie (born 27 June 1953) is an American statistician and computer scientist. He is currently serving as the John A. Overdeck Professor of Mathematical Sciences and Professor of Statistics at Stanford University. Hastie is known for ...
,
Iain Johnstone Iain Johnstone (born 8 April 1943 in Reading, Berkshire) is an English author, broadcaster and television producer. Early life Johnstone attended Crosfields School, Reading, Campbell College, Belfast and Bristol University. Career Johnst ...
and
Robert Tibshirani Robert Tibshirani (born July 10, 1956) is a professor in the Departments of Statistics and Biomedical Data Science at Stanford University. He was a professor at the University of Toronto from 1985 to 1998. In his work, he develops statistical to ...
. Suppose we expect a response variable to be determined by a linear combination of a subset of potential covariates. Then the LARS algorithm provides a means of producing an estimate of which variables to include, as well as their coefficients. Instead of giving a vector result, the LARS solution consists of a curve denoting the solution for each value of the
L1 norm In mathematics, the spaces are function spaces defined using a natural generalization of the -norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue , although according to the Bourbaki ...
of the parameter vector. The algorithm is similar to forward
stepwise regression In statistics, stepwise regression is a method of fitting regression models in which the choice of predictive variables is carried out by an automatic procedure. In each step, a variable is considered for addition to or subtraction from the set of ...
, but instead of including variables at each step, the estimated parameters are increased in a direction equiangular to each one's correlations with the residual.


Pros and cons

The advantages of the LARS method are: # It is computationally just as fast as forward selection. # It produces a full piecewise linear solution path, which is useful in cross-validation or similar attempts to tune the model. # If two variables are almost equally correlated with the response, then their coefficients should increase at approximately the same rate. The algorithm thus behaves as intuition would expect, and also is more stable. # It is easily modified to produce efficient algorithms for other methods producing similar results, like the
lasso A lasso ( or ), also called lariat, riata, or reata (all from Castilian, la reata 're-tied rope'), is a loop of rope designed as a restraint to be thrown around a target and tightened when pulled. It is a well-known tool of the Spanish an ...
and forward stagewise regression. # It is effective in contexts where ''p'' >> ''n'' (i.e., when the number of predictors ''p'' is significantly greater than the number of points ''n'') The disadvantages of the LARS method include: # With any amount of noise in the dependent variable and with high dimensional multicollinear independent variables, there is no reason to believe that the selected variables will have a high probability of being the actual underlying causal variables. This problem is not unique to LARS, as it is a general problem with variable selection approaches that seek to find underlying deterministic components. Yet, because LARS is based upon an iterative refitting of the residuals, it would appear to be especially sensitive to the effects of noise. This problem is discussed in detail by Weisberg in the discussion section of the Efron et al. (2004) Annals of Statistics article. Weisberg provides an empirical example based upon re-analysis of data originally used to validate LARS that the variable selection appears to have problems with highly correlated variables. # Since almost all high dimensional data in the real world will just by chance exhibit some fair degree of collinearity across at least some variables, the problem that LARS has with correlated variables may limit its application to high dimensional data.


Algorithm

The basic steps of the Least-angle regression algorithm are: * Start with all coefficients \beta equal to zero. * Find the predictor x_j most correlated with y. * Increase the coefficient \beta_j in the direction of the sign of its correlation with y. Take residuals r = y - \hat along the way. Stop when some other predictor x_k has as much correlation with r as x_j has. * Increase (\beta_j, \beta_k) in their joint least squares direction, until some other predictor x_m has as much correlation with the residual r. * Increase (\beta_j, \beta_k, \beta_m) in their joint least squares direction, until some other predictor x_n has as much correlation with the residual r. * Continue until: all predictors are in the model.


Software implementation

Least-angle regression is implemented in R via th
lars
package, in Python with th
scikit-learn
package, and in SAS via th
GLMSELECT
procedure.


See also

*
High-dimensional statistics In statistical theory, the field of high-dimensional statistics studies data whose dimension is larger than typically considered in classical multivariate analysis. The area arose owing to the emergence of many modern data sets in which the dimensi ...
*
Lasso (statistics) In statistics and machine learning, lasso (least absolute shrinkage and selection operator; also Lasso or LASSO) is a regression analysis method that performs both variable selection and regularization in order to enhance the prediction accuracy ...
*
Regression analysis In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome' or 'response' variable, or a 'label' in machine learning parlance) and one ...
*
Model selection Model selection is the task of selecting a statistical model from a set of candidate models, given data. In the simplest cases, a pre-existing set of data is considered. However, the task can also involve the design of experiments such that the ...


References

{{DEFAULTSORT:Least-Angle Regression Estimation theory Parametric statistics Regression variable selection Single-equation methods (econometrics)