Mallows's Cp
   HOME

TheInfoList



OR:

In statistics, Mallows's ''Cp'', named for Colin Lingwood Mallows, is used to assess the fit of a
regression model In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome' or 'response' variable, or a 'label' in machine learning parlance) and one ...
that has been estimated using
ordinary least squares In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model (with fixed level-one effects of a linear function of a set of explanatory variables) by the ...
. It is applied in the context of
model selection Model selection is the task of selecting a statistical model from a set of candidate models, given data. In the simplest cases, a pre-existing set of data is considered. However, the task can also involve the design of experiments such that the ...
, where a number of predictor variables are available for predicting some outcome, and the goal is to find the best model involving a subset of these predictors. A small value of Cp means that the model is relatively precise. Mallows's ''Cp'' has been shown to be equivalent to
Akaike information criterion The Akaike information criterion (AIC) is an estimator of prediction error and thereby relative quality of statistical models for a given set of data. Given a collection of models for the data, AIC estimates the quality of each model, relative to e ...
in the special case of Gaussian linear regression.


Definition and properties

Mallows's ''Cp'' addresses the issue of overfitting, in which model selection statistics such as the residual sum of squares always get smaller as more variables are added to a model. Thus, if we aim to select the model giving the smallest residual sum of squares, the model including all variables would always be selected. Instead, the ''Cp'' statistic calculated on a sample of data estimates the sum squared prediction error (SSPE) as its
population Population typically refers to the number of people in a single area, whether it be a city or town, region, country, continent, or the world. Governments typically quantify the size of the resident population within their jurisdiction using a ...
target : E\sum_i (\hat_i - E(Y_i\mid X_i))^2/\sigma^2, where \hat_i is the fitted value from the regression model for the ''i''th case, ''E''(''Y''''i'' ,  ''X''''i'') is the expected value for the ''i''th case, and σ2 is the error variance (assumed constant across the cases). The MSPE will not automatically get smaller as more variables are added. The optimum model under this criterion is a compromise influenced by the sample size, the
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
s of the different predictors, and the degree of
collinearity In geometry, collinearity of a set of points is the property of their lying on a single line. A set of points with this property is said to be collinear (sometimes spelled as colinear). In greater generality, the term has been used for aligned o ...
between them. If ''P'' regressors are selected from a set of ''K'' > ''P'', the ''Cp'' statistic for that particular set of regressors is defined as: : C_p= - N + 2(P+1), where *SSE_p = \sum_^N(Y_i-Y_)^2 is the error sum of squares for the model with ''P'' regressors, *''Y''pi is the
predict A prediction (Latin ''præ-'', "before," and ''dicere'', "to say"), or forecast, is a statement about a future event or data. They are often, but not always, based upon experience or knowledge. There is no universal agreement about the exact ...
ed value of the ''i''th observation of ''Y'' from the ''P'' regressors, *''S''2 is the residual mean square after regression on the complete set of ''K'' regressors and can be estimated by mean square error ''MSE'', * and ''N'' is the
sample size Sample size determination is the act of choosing the number of observations or replicates to include in a statistical sample. The sample size is an important feature of any empirical study in which the goal is to make inferences about a populatio ...
.


Alternative definition

Given a linear model such as: : Y = \beta_0 + \beta_1X_1+\cdots+\beta_pX_p + \varepsilon where: * \beta_0,\ldots,\beta_p are coefficients for predictor variables X_1,\ldots,X_p * \varepsilon represents error An alternate version of ''Cp'' can also be defined as: : C_p=\frac(\operatorname + 2p\hat^2) where * RSS is the residual sum of squares on a training set of data * is the number of predictors * and \hat^2 refers to an estimate of the variance associated with each response in the linear model (estimated on a model containing all predictors) Note that this version of the ''Cp'' does not give equivalent values to the earlier version, but the model with the smallest ''Cp'' from this definition will also be the same model with the smallest ''Cp'' from the earlier definition.


Limitations

The ''Cp'' criterion suffers from two main limitationsGiraud, C. (2015), ''Introduction to high-dimensional statistics'', Chapman & Hall/CRC, # the ''Cp'' approximation is only valid for large sample size; # the ''Cp'' cannot handle complex collections of models as in the variable selection (or
feature selection In machine learning and statistics, feature selection, also known as variable selection, attribute selection or variable subset selection, is the process of selecting a subset of relevant features (variables, predictors) for use in model construc ...
) problem.


Practical use

The ''Cp'' statistic is often used as a stopping rule for various forms of stepwise regression. Mallows proposed the statistic as a criterion for selecting among many alternative subset regressions. Under a model not suffering from appreciable lack of fit (bias), ''Cp'' has expectation nearly equal to ''P''; otherwise the expectation is roughly ''P'' plus a positive bias term. Nevertheless, even though it has expectation greater than or equal to ''P'', there is nothing to prevent ''Cp'' < ''P'' or even ''Cp'' < 0 in extreme cases. It is suggested that one should choose a subset that has ''Cp'' approaching ''P'', from above, for a list of subsets ordered by increasing ''P''. In practice, the positive bias can be adjusted for by selecting a model from the ordered list of subsets, such that ''Cp'' < 2''P''. Since the sample-based ''Cp'' statistic is an estimate of the MSPE, using ''Cp'' for model selection does not completely guard against overfitting. For instance, it is possible that the selected model will be one in which the sample ''Cp'' was a particularly severe underestimate of the MSPE. Model selection statistics such as ''Cp'' are generally not used blindly, but rather information about the field of application, the intended use of the model, and any known biases in the data are taken into account in the process of model selection.


See also

* Goodness of fit: Regression analysis * Coefficient of determination


References


Further reading

* * * {{cite book , last1=Judge , first1=George G. , first2=William E. , last2=Griffiths , first3=R. Carter , last3=Hill , first4=Tsoung-Chao , last4=Lee , year=1980 , title=The Theory and Practice of Econometrics , location= New York , publisher=Wiley , pages=417–423 , isbn=978-0-471-05938-7 Regression diagnostics Regression variable selection